var/home/core/zuul-output/0000755000175000017500000000000015113536126014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113547266015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004705670715113547260017721 0ustar rootrootDec 02 10:19:05 crc systemd[1]: Starting Kubernetes Kubelet... Dec 02 10:19:05 crc restorecon[4678]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:05 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 10:19:06 crc restorecon[4678]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 02 10:19:06 crc kubenswrapper[4679]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 10:19:06 crc kubenswrapper[4679]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 02 10:19:06 crc kubenswrapper[4679]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 10:19:06 crc kubenswrapper[4679]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 10:19:06 crc kubenswrapper[4679]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 02 10:19:06 crc kubenswrapper[4679]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.772195 4679 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774886 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774903 4679 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774908 4679 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774912 4679 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774916 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774920 4679 feature_gate.go:330] unrecognized feature gate: Example Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774924 4679 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774929 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774934 4679 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774937 4679 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774942 4679 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774946 4679 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774950 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774953 4679 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774957 4679 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774960 4679 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774964 4679 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774967 4679 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774971 4679 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774974 4679 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774978 4679 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774981 4679 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774985 4679 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774988 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774992 4679 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774995 4679 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.774998 4679 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775007 4679 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775011 4679 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775014 4679 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775018 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775022 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775025 4679 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775029 4679 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775033 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775037 4679 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775043 4679 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775049 4679 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775054 4679 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775059 4679 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775064 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775069 4679 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775076 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775082 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775088 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775092 4679 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775097 4679 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775100 4679 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775105 4679 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775109 4679 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775113 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775117 4679 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775121 4679 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775124 4679 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775128 4679 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775131 4679 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775135 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775138 4679 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775141 4679 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775145 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775149 4679 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775154 4679 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775158 4679 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775162 4679 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775166 4679 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775170 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775173 4679 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775177 4679 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775181 4679 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775185 4679 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.775190 4679 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775584 4679 flags.go:64] FLAG: --address="0.0.0.0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775596 4679 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775604 4679 flags.go:64] FLAG: --anonymous-auth="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775609 4679 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775615 4679 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775619 4679 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775625 4679 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775630 4679 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775634 4679 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775638 4679 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775642 4679 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775647 4679 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775651 4679 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775655 4679 flags.go:64] FLAG: --cgroup-root="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775659 4679 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775663 4679 flags.go:64] FLAG: --client-ca-file="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775667 4679 flags.go:64] FLAG: --cloud-config="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775670 4679 flags.go:64] FLAG: --cloud-provider="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775674 4679 flags.go:64] FLAG: --cluster-dns="[]" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775679 4679 flags.go:64] FLAG: --cluster-domain="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775683 4679 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775687 4679 flags.go:64] FLAG: --config-dir="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775691 4679 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775695 4679 flags.go:64] FLAG: --container-log-max-files="5" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775702 4679 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775706 4679 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775710 4679 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775714 4679 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775719 4679 flags.go:64] FLAG: --contention-profiling="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775723 4679 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775727 4679 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775731 4679 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775735 4679 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775740 4679 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775744 4679 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775749 4679 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775753 4679 flags.go:64] FLAG: --enable-load-reader="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775758 4679 flags.go:64] FLAG: --enable-server="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775762 4679 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775771 4679 flags.go:64] FLAG: --event-burst="100" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775776 4679 flags.go:64] FLAG: --event-qps="50" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775780 4679 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775784 4679 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775788 4679 flags.go:64] FLAG: --eviction-hard="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775793 4679 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775797 4679 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775800 4679 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775805 4679 flags.go:64] FLAG: --eviction-soft="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775809 4679 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775812 4679 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775816 4679 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775821 4679 flags.go:64] FLAG: --experimental-mounter-path="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775824 4679 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775828 4679 flags.go:64] FLAG: --fail-swap-on="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775832 4679 flags.go:64] FLAG: --feature-gates="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775837 4679 flags.go:64] FLAG: --file-check-frequency="20s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775842 4679 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775846 4679 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775850 4679 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775854 4679 flags.go:64] FLAG: --healthz-port="10248" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775858 4679 flags.go:64] FLAG: --help="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775862 4679 flags.go:64] FLAG: --hostname-override="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775866 4679 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775870 4679 flags.go:64] FLAG: --http-check-frequency="20s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775874 4679 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775878 4679 flags.go:64] FLAG: --image-credential-provider-config="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775881 4679 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775886 4679 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775890 4679 flags.go:64] FLAG: --image-service-endpoint="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775894 4679 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775898 4679 flags.go:64] FLAG: --kube-api-burst="100" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775902 4679 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775907 4679 flags.go:64] FLAG: --kube-api-qps="50" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775911 4679 flags.go:64] FLAG: --kube-reserved="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775915 4679 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775919 4679 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775923 4679 flags.go:64] FLAG: --kubelet-cgroups="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775927 4679 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775931 4679 flags.go:64] FLAG: --lock-file="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775935 4679 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775939 4679 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775943 4679 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775949 4679 flags.go:64] FLAG: --log-json-split-stream="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775953 4679 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775957 4679 flags.go:64] FLAG: --log-text-split-stream="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775961 4679 flags.go:64] FLAG: --logging-format="text" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775965 4679 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775969 4679 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775973 4679 flags.go:64] FLAG: --manifest-url="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775977 4679 flags.go:64] FLAG: --manifest-url-header="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775982 4679 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775986 4679 flags.go:64] FLAG: --max-open-files="1000000" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775991 4679 flags.go:64] FLAG: --max-pods="110" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775995 4679 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.775999 4679 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776003 4679 flags.go:64] FLAG: --memory-manager-policy="None" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776006 4679 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776010 4679 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776014 4679 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776019 4679 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776027 4679 flags.go:64] FLAG: --node-status-max-images="50" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776031 4679 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776036 4679 flags.go:64] FLAG: --oom-score-adj="-999" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776040 4679 flags.go:64] FLAG: --pod-cidr="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776044 4679 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776050 4679 flags.go:64] FLAG: --pod-manifest-path="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776054 4679 flags.go:64] FLAG: --pod-max-pids="-1" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776058 4679 flags.go:64] FLAG: --pods-per-core="0" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776063 4679 flags.go:64] FLAG: --port="10250" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776067 4679 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776071 4679 flags.go:64] FLAG: --provider-id="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776075 4679 flags.go:64] FLAG: --qos-reserved="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776079 4679 flags.go:64] FLAG: --read-only-port="10255" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776084 4679 flags.go:64] FLAG: --register-node="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776088 4679 flags.go:64] FLAG: --register-schedulable="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776091 4679 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776098 4679 flags.go:64] FLAG: --registry-burst="10" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776102 4679 flags.go:64] FLAG: --registry-qps="5" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776107 4679 flags.go:64] FLAG: --reserved-cpus="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776111 4679 flags.go:64] FLAG: --reserved-memory="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776116 4679 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776120 4679 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776124 4679 flags.go:64] FLAG: --rotate-certificates="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776129 4679 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776132 4679 flags.go:64] FLAG: --runonce="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776136 4679 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776140 4679 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776145 4679 flags.go:64] FLAG: --seccomp-default="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776149 4679 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776153 4679 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776157 4679 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776161 4679 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776165 4679 flags.go:64] FLAG: --storage-driver-password="root" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776169 4679 flags.go:64] FLAG: --storage-driver-secure="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776174 4679 flags.go:64] FLAG: --storage-driver-table="stats" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776178 4679 flags.go:64] FLAG: --storage-driver-user="root" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776182 4679 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776186 4679 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776190 4679 flags.go:64] FLAG: --system-cgroups="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776195 4679 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776201 4679 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776205 4679 flags.go:64] FLAG: --tls-cert-file="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776210 4679 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776214 4679 flags.go:64] FLAG: --tls-min-version="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776219 4679 flags.go:64] FLAG: --tls-private-key-file="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776223 4679 flags.go:64] FLAG: --topology-manager-policy="none" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776227 4679 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776231 4679 flags.go:64] FLAG: --topology-manager-scope="container" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776235 4679 flags.go:64] FLAG: --v="2" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776240 4679 flags.go:64] FLAG: --version="false" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776246 4679 flags.go:64] FLAG: --vmodule="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776251 4679 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776256 4679 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776364 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776369 4679 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776373 4679 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776377 4679 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776380 4679 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776384 4679 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776387 4679 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776391 4679 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776394 4679 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776398 4679 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776402 4679 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776405 4679 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776409 4679 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776412 4679 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776416 4679 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776419 4679 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776422 4679 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776426 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776429 4679 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776433 4679 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776436 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776440 4679 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776443 4679 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776447 4679 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776451 4679 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776455 4679 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776460 4679 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776464 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776469 4679 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776473 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776477 4679 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776480 4679 feature_gate.go:330] unrecognized feature gate: Example Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776484 4679 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776488 4679 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776492 4679 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776496 4679 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776500 4679 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776508 4679 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776514 4679 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776519 4679 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776522 4679 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776526 4679 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776530 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776533 4679 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776537 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776541 4679 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776544 4679 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776547 4679 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776551 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776555 4679 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776558 4679 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776562 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776565 4679 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776569 4679 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776572 4679 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776576 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776579 4679 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776583 4679 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776587 4679 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776593 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776597 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776601 4679 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776605 4679 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776609 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776614 4679 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776618 4679 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776621 4679 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776625 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776628 4679 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776634 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.776637 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.776648 4679 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.784002 4679 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.784033 4679 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784118 4679 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784125 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784130 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784133 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784137 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784141 4679 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784145 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784149 4679 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784153 4679 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784156 4679 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784160 4679 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784164 4679 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784167 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784171 4679 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784174 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784177 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784181 4679 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784184 4679 feature_gate.go:330] unrecognized feature gate: Example Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784188 4679 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784192 4679 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784195 4679 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784199 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784202 4679 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784206 4679 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784209 4679 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784214 4679 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784220 4679 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784224 4679 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784228 4679 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784232 4679 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784235 4679 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784242 4679 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784245 4679 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784249 4679 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784254 4679 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784259 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784263 4679 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784267 4679 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784271 4679 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784274 4679 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784278 4679 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784282 4679 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784285 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784289 4679 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784292 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784296 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784315 4679 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784320 4679 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784324 4679 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784327 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784331 4679 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784335 4679 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784340 4679 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784344 4679 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784347 4679 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784351 4679 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784354 4679 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784357 4679 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784361 4679 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784364 4679 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784368 4679 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784371 4679 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784375 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784380 4679 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784384 4679 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784388 4679 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784396 4679 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784404 4679 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784409 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784413 4679 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784421 4679 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.784429 4679 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784552 4679 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784560 4679 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784564 4679 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784567 4679 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784571 4679 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784576 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784580 4679 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784584 4679 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784588 4679 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784592 4679 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784596 4679 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784600 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784604 4679 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784607 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784611 4679 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784615 4679 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784619 4679 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784623 4679 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784627 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784630 4679 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784634 4679 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784638 4679 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784641 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784645 4679 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784649 4679 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784652 4679 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784656 4679 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784659 4679 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784662 4679 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784666 4679 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784671 4679 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784676 4679 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784680 4679 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784684 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784689 4679 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784693 4679 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784696 4679 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784701 4679 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784704 4679 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784708 4679 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784711 4679 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784714 4679 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784718 4679 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784722 4679 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784725 4679 feature_gate.go:330] unrecognized feature gate: Example Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784729 4679 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784732 4679 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784736 4679 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784739 4679 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784742 4679 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784746 4679 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784749 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784753 4679 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784756 4679 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784761 4679 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784765 4679 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784768 4679 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784772 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784775 4679 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784779 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784782 4679 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784785 4679 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784789 4679 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784792 4679 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784796 4679 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784799 4679 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784802 4679 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784806 4679 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784809 4679 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784812 4679 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.784818 4679 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.784823 4679 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.785232 4679 server.go:940] "Client rotation is on, will bootstrap in background" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.787443 4679 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.787716 4679 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.788263 4679 server.go:997] "Starting client certificate rotation" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.788287 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.788853 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-19 05:24:15.41353508 +0000 UTC Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.788947 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.794143 4679 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.798026 4679 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.798058 4679 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.804908 4679 log.go:25] "Validated CRI v1 runtime API" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.819001 4679 log.go:25] "Validated CRI v1 image API" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.820217 4679 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.822279 4679 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-02-10-14-45-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.822324 4679 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.835346 4679 manager.go:217] Machine: {Timestamp:2025-12-02 10:19:06.834382666 +0000 UTC m=+0.164521546 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f BootID:128cffb4-d188-4793-ba88-90c1cc2d9356 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f6:ab:e8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f6:ab:e8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:94:3c:16 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:36:da:8f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:75:f5:7f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:aa:0d:68 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1e:dc:e9:97:02:70 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1a:32:4e:fd:96:33 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.835625 4679 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.835749 4679 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836018 4679 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836184 4679 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836219 4679 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836433 4679 topology_manager.go:138] "Creating topology manager with none policy" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836444 4679 container_manager_linux.go:303] "Creating device plugin manager" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836622 4679 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836636 4679 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836801 4679 state_mem.go:36] "Initialized new in-memory state store" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.836866 4679 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.837500 4679 kubelet.go:418] "Attempting to sync node with API server" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.837522 4679 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.837535 4679 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.837545 4679 kubelet.go:324] "Adding apiserver pod source" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.837556 4679 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.839768 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.839769 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.839851 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.839877 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.839936 4679 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.840242 4679 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.840835 4679 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841438 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841484 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841503 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841513 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841531 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841542 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841557 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841572 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841579 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841587 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841596 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841603 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.841804 4679 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.848133 4679 server.go:1280] "Started kubelet" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.849529 4679 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.850613 4679 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.850621 4679 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.851619 4679 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 02 10:19:06 crc systemd[1]: Started Kubernetes Kubelet. Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.852768 4679 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d5eb1d366db09 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 10:19:06.848058121 +0000 UTC m=+0.178197251,LastTimestamp:2025-12-02 10:19:06.848058121 +0000 UTC m=+0.178197251,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.856267 4679 server.go:460] "Adding debug handlers to kubelet server" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860199 4679 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860247 4679 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860293 4679 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 08:53:02.18501096 +0000 UTC Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860370 4679 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 718h33m55.32464348s for next certificate rotation Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.860653 4679 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860687 4679 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860675 4679 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.860718 4679 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.861068 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="200ms" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.861221 4679 factory.go:55] Registering systemd factory Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.861246 4679 factory.go:221] Registration of the systemd container factory successfully Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.861347 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.861565 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.864061 4679 factory.go:153] Registering CRI-O factory Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.864086 4679 factory.go:221] Registration of the crio container factory successfully Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.864163 4679 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.864190 4679 factory.go:103] Registering Raw factory Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.864208 4679 manager.go:1196] Started watching for new ooms in manager Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.864969 4679 manager.go:319] Starting recovery of all containers Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871355 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871396 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871407 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871448 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871458 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871467 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871500 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871510 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871574 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871584 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871593 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871601 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871610 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871621 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871629 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871637 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871663 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871690 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871699 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871707 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871716 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871744 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871753 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871761 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871787 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871796 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871806 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871817 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871859 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871868 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871876 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871885 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871911 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871943 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871952 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871960 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871970 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871978 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871986 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.871994 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872037 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872046 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872095 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872104 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872112 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872121 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872129 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.872138 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873275 4679 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873321 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873343 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873358 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873397 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873447 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873463 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873474 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873551 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873565 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873576 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873585 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873604 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873618 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873631 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873693 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873789 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873802 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873812 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873832 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873866 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873880 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873891 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873902 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873934 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873945 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873956 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873967 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.873978 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874014 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874073 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874088 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874131 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874143 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874154 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874166 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874203 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874215 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874226 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874238 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874268 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874280 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874346 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874357 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874369 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874382 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874394 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874406 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874487 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874499 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874509 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874530 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874539 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874548 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874556 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874564 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874588 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874645 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874667 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874676 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874686 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874695 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874704 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874713 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874739 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874748 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874785 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874796 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874818 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874828 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874865 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874879 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874908 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874917 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874925 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874933 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874942 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874950 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874959 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.874966 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875002 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875043 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875052 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875060 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875081 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875089 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875097 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875105 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875130 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875139 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875148 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875156 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875165 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875174 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875193 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875207 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875221 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875507 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875521 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875538 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875575 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875587 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875599 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875610 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875622 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875634 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875644 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875655 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875666 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875676 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875686 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875696 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875736 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875749 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875762 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875775 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875785 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875794 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875802 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875810 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875818 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875828 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875845 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875853 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875867 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875875 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875883 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875892 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875901 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875909 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875922 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875930 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875938 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875946 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875955 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875963 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875972 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875980 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875989 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.875998 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876007 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876016 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876026 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876036 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876045 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876053 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876061 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876069 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876078 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876087 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876096 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876104 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876113 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876123 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876131 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876139 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876147 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876156 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876166 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876177 4679 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876189 4679 reconstruct.go:97] "Volume reconstruction finished" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.876198 4679 reconciler.go:26] "Reconciler: start to sync state" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.887338 4679 manager.go:324] Recovery completed Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.898446 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.900819 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.900852 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.900865 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.902217 4679 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.902233 4679 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.902326 4679 state_mem.go:36] "Initialized new in-memory state store" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.904220 4679 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.907247 4679 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.907364 4679 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.907414 4679 kubelet.go:2335] "Starting kubelet main sync loop" Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.907506 4679 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 02 10:19:06 crc kubenswrapper[4679]: W1202 10:19:06.907914 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.907958 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.916217 4679 policy_none.go:49] "None policy: Start" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.917224 4679 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.917252 4679 state_mem.go:35] "Initializing new in-memory state store" Dec 02 10:19:06 crc kubenswrapper[4679]: E1202 10:19:06.961163 4679 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.994659 4679 manager.go:334] "Starting Device Plugin manager" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.995447 4679 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.995462 4679 server.go:79] "Starting device plugin registration server" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.995789 4679 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.995804 4679 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.996038 4679 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.996105 4679 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 02 10:19:06 crc kubenswrapper[4679]: I1202 10:19:06.996113 4679 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.002250 4679 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.008419 4679 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.008518 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.009399 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.009431 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.009442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.009577 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.009846 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.009895 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010355 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010393 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010405 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010468 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010602 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010670 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010868 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010902 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.010911 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011068 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011084 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011092 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011195 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011274 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011298 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011324 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011395 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.011446 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.012112 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.012170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.012190 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.012409 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.013298 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.013369 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.014756 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.014784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.014793 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.014966 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.014996 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015146 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015184 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015227 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015247 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015258 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015658 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015681 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.015689 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.061727 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="400ms" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078478 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078549 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078667 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078739 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078772 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078822 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078849 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078876 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078903 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078928 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078957 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.078983 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.079015 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.079058 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.079088 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.096458 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.098023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.098058 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.098069 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.098092 4679 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.098592 4679 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180166 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180227 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180253 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180281 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180299 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180328 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180342 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180359 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180390 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180409 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180426 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180442 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180457 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180473 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180456 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180519 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180541 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180577 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180459 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180490 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180481 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180636 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180692 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180618 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180643 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180599 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180652 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180548 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180696 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.180540 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.299599 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.300693 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.300733 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.300742 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.300763 4679 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.301135 4679 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.347020 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.355335 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.376096 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.384488 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-47ea047765f95e72d02d21ce8dcdc5c4eb743b274d8d0ce23e1a719ebe8dca6a WatchSource:0}: Error finding container 47ea047765f95e72d02d21ce8dcdc5c4eb743b274d8d0ce23e1a719ebe8dca6a: Status 404 returned error can't find the container with id 47ea047765f95e72d02d21ce8dcdc5c4eb743b274d8d0ce23e1a719ebe8dca6a Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.385920 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-7a3a5f69d3336dacc651118c23e79547ca5f8d9a803704df337c04e63085e9de WatchSource:0}: Error finding container 7a3a5f69d3336dacc651118c23e79547ca5f8d9a803704df337c04e63085e9de: Status 404 returned error can't find the container with id 7a3a5f69d3336dacc651118c23e79547ca5f8d9a803704df337c04e63085e9de Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.389276 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.390646 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f44084839fbb26f7bec4432c028381beced85c475a4123fdd0af6a59d64e0c63 WatchSource:0}: Error finding container f44084839fbb26f7bec4432c028381beced85c475a4123fdd0af6a59d64e0c63: Status 404 returned error can't find the container with id f44084839fbb26f7bec4432c028381beced85c475a4123fdd0af6a59d64e0c63 Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.395658 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.400635 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-602c756c0b8cd182d6924ec82bc5d39adc6b137ad5c4e64536f6d0bdb081a793 WatchSource:0}: Error finding container 602c756c0b8cd182d6924ec82bc5d39adc6b137ad5c4e64536f6d0bdb081a793: Status 404 returned error can't find the container with id 602c756c0b8cd182d6924ec82bc5d39adc6b137ad5c4e64536f6d0bdb081a793 Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.414487 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-2abe31bb3716e2bcabecd6426842ac11bc779ca5070c6264a746b1957531d216 WatchSource:0}: Error finding container 2abe31bb3716e2bcabecd6426842ac11bc779ca5070c6264a746b1957531d216: Status 404 returned error can't find the container with id 2abe31bb3716e2bcabecd6426842ac11bc779ca5070c6264a746b1957531d216 Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.462484 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="800ms" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.702168 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.703076 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.703112 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.703120 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.703141 4679 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.703514 4679 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.850950 4679 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.901500 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.901589 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.914473 4679 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb" exitCode=0 Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.914557 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.914660 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f44084839fbb26f7bec4432c028381beced85c475a4123fdd0af6a59d64e0c63"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.914757 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.916597 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.916647 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.916662 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.917266 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d" exitCode=0 Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.917293 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.917353 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"47ea047765f95e72d02d21ce8dcdc5c4eb743b274d8d0ce23e1a719ebe8dca6a"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.917446 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.918416 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.918498 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.918554 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.919879 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.919914 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7a3a5f69d3336dacc651118c23e79547ca5f8d9a803704df337c04e63085e9de"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.920785 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.921499 4679 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6cd659014febc043915eca2d111c71cebbd37107f2c5c02328490b017d112165" exitCode=0 Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.921558 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6cd659014febc043915eca2d111c71cebbd37107f2c5c02328490b017d112165"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.921581 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2abe31bb3716e2bcabecd6426842ac11bc779ca5070c6264a746b1957531d216"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.921685 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.924709 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.924742 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.924752 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.924714 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.924785 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.924797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.926383 4679 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6" exitCode=0 Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.926421 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.926446 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"602c756c0b8cd182d6924ec82bc5d39adc6b137ad5c4e64536f6d0bdb081a793"} Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.926518 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.927358 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.927416 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:07 crc kubenswrapper[4679]: I1202 10:19:07.927428 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:07 crc kubenswrapper[4679]: W1202 10:19:07.994472 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:07 crc kubenswrapper[4679]: E1202 10:19:07.994547 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:08 crc kubenswrapper[4679]: W1202 10:19:08.014195 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:08 crc kubenswrapper[4679]: E1202 10:19:08.014272 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:08 crc kubenswrapper[4679]: W1202 10:19:08.140844 4679 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 02 10:19:08 crc kubenswrapper[4679]: E1202 10:19:08.140935 4679 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 02 10:19:08 crc kubenswrapper[4679]: E1202 10:19:08.263564 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="1.6s" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.504209 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.505115 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.505149 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.505159 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.505181 4679 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 10:19:08 crc kubenswrapper[4679]: E1202 10:19:08.505454 4679 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.877564 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.931624 4679 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4fde56319ce232d90d9914492d695f620b7bad41caf8e92407d8ae7e18e19a82" exitCode=0 Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.931683 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4fde56319ce232d90d9914492d695f620b7bad41caf8e92407d8ae7e18e19a82"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.931784 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.932479 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.932500 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.932508 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.933742 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.933792 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.934859 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.934902 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.934915 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.937483 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.937529 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.937547 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.937618 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.938561 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.938585 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.938595 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.945593 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.945633 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.945647 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.945660 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.952144 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.952178 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.952189 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e"} Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.952236 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.952990 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.953016 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:08 crc kubenswrapper[4679]: I1202 10:19:08.953032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.960202 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f"} Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.960556 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.961877 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.961937 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.961961 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.962789 4679 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8d30ed1d5974ce944aaae6a637f4cfb039745c28418f2d74892b4e6b04760117" exitCode=0 Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.962859 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8d30ed1d5974ce944aaae6a637f4cfb039745c28418f2d74892b4e6b04760117"} Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.962904 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.962946 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.963038 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.963101 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964518 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964622 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964844 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964899 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964926 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.964981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:09 crc kubenswrapper[4679]: I1202 10:19:09.965006 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.106346 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.107371 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.107411 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.107419 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.107444 4679 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970510 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c8017c28f94ddceb93c72ae06a24f504db8794697cfb8f8f9fd6367ad0155832"} Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970564 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0ae0c1af462298c0b133d1327e71ea81a1fa4ad11d6d38e6a0ae1dad71492377"} Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970582 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb734943dfafcce8dabc7bf4cd380ffece43867d0d0693b84ede53df907ffd1e"} Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970593 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970737 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970599 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c861a3ffb113903b6efee9a112f2c6343e06a1417cb3de8113c2238b31aad042"} Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970797 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b14a45bcb5c124115f0eafaba65af26d4940ea50fa5b5ce626b3878e787cc96e"} Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.970843 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.972559 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.972594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.972609 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.972634 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.972675 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:10 crc kubenswrapper[4679]: I1202 10:19:10.972696 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.044138 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.270567 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.390140 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.972994 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.973004 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.973164 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.973697 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.973722 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.973733 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.973988 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.974035 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:11 crc kubenswrapper[4679]: I1202 10:19:11.974046 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.059966 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.216872 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.217074 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.218330 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.218374 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.218385 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.975200 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.975345 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.976511 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.976578 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.976601 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.976786 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.976836 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:12 crc kubenswrapper[4679]: I1202 10:19:12.976852 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.672131 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.672350 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.674696 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.674752 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.674764 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.815912 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.978991 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.979033 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.980617 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.980669 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.980690 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.981640 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.981701 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:13 crc kubenswrapper[4679]: I1202 10:19:13.981721 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.170773 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.170970 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.172163 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.172233 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.172249 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.178639 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.980783 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.981692 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.981710 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:14 crc kubenswrapper[4679]: I1202 10:19:14.981719 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:15 crc kubenswrapper[4679]: I1202 10:19:15.475518 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:15 crc kubenswrapper[4679]: I1202 10:19:15.982984 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:15 crc kubenswrapper[4679]: I1202 10:19:15.984030 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:15 crc kubenswrapper[4679]: I1202 10:19:15.984076 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:15 crc kubenswrapper[4679]: I1202 10:19:15.984090 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:17 crc kubenswrapper[4679]: E1202 10:19:17.002500 4679 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.787281 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.787862 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.789442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.789485 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.789496 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.793028 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.851673 4679 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 02 10:19:18 crc kubenswrapper[4679]: E1202 10:19:18.879536 4679 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.990947 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.991775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.991810 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:18 crc kubenswrapper[4679]: I1202 10:19:18.991820 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:19 crc kubenswrapper[4679]: I1202 10:19:19.215553 4679 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 02 10:19:19 crc kubenswrapper[4679]: I1202 10:19:19.215623 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 02 10:19:19 crc kubenswrapper[4679]: I1202 10:19:19.220047 4679 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 02 10:19:19 crc kubenswrapper[4679]: I1202 10:19:19.220218 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.398819 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.399588 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.400733 4679 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.400809 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.400854 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.400878 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.400832 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.406082 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.648098 4679 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.648160 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.788139 4679 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.788230 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.998266 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.999091 4679 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.999153 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.999483 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.999531 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:21 crc kubenswrapper[4679]: I1202 10:19:21.999558 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:22 crc kubenswrapper[4679]: I1202 10:19:22.061183 4679 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 10:19:22 crc kubenswrapper[4679]: I1202 10:19:22.061248 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 10:19:22 crc kubenswrapper[4679]: I1202 10:19:22.916137 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 02 10:19:22 crc kubenswrapper[4679]: I1202 10:19:22.926555 4679 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 02 10:19:22 crc kubenswrapper[4679]: I1202 10:19:22.941110 4679 csr.go:261] certificate signing request csr-tml76 is approved, waiting to be issued Dec 02 10:19:22 crc kubenswrapper[4679]: I1202 10:19:22.979857 4679 csr.go:257] certificate signing request csr-tml76 is issued Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.850284 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.850494 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.851546 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.851574 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.851586 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.875071 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.980897 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-02 10:14:22 +0000 UTC, rotation deadline is 2026-08-24 19:12:13.333978141 +0000 UTC Dec 02 10:19:23 crc kubenswrapper[4679]: I1202 10:19:23.980963 4679 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6368h52m49.353017553s for next certificate rotation Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.002531 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.003405 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.003433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.003442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:24 crc kubenswrapper[4679]: E1202 10:19:24.215801 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.216913 4679 trace.go:236] Trace[164762822]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 10:19:09.779) (total time: 14437ms): Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[164762822]: ---"Objects listed" error: 14437ms (10:19:24.216) Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[164762822]: [14.437163879s] [14.437163879s] END Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.216937 4679 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.217241 4679 trace.go:236] Trace[826550250]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 10:19:11.053) (total time: 13163ms): Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[826550250]: ---"Objects listed" error: 13163ms (10:19:24.217) Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[826550250]: [13.16381904s] [13.16381904s] END Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.217289 4679 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.219470 4679 trace.go:236] Trace[1913189234]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 10:19:10.440) (total time: 13778ms): Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[1913189234]: ---"Objects listed" error: 13778ms (10:19:24.219) Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[1913189234]: [13.778744737s] [13.778744737s] END Dec 02 10:19:24 crc kubenswrapper[4679]: E1202 10:19:24.219488 4679 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.219496 4679 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.220249 4679 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.221782 4679 trace.go:236] Trace[13835431]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 10:19:10.459) (total time: 13762ms): Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[13835431]: ---"Objects listed" error: 13762ms (10:19:24.221) Dec 02 10:19:24 crc kubenswrapper[4679]: Trace[13835431]: [13.762275341s] [13.762275341s] END Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.221797 4679 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.850826 4679 apiserver.go:52] "Watching apiserver" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.936389 4679 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.936794 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-multus/multus-r5sft","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-multus/multus-additional-cni-plugins-2lmfz","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-7pw64","openshift-machine-config-operator/machine-config-daemon-lzf8q"] Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.937125 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.937220 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.937259 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:24 crc kubenswrapper[4679]: E1202 10:19:24.937377 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:24 crc kubenswrapper[4679]: E1202 10:19:24.937372 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.937580 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.937597 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.937804 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:24 crc kubenswrapper[4679]: E1202 10:19:24.937937 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.938203 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-r5sft" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.938711 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.938807 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.938867 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.960221 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.962343 4679 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.963412 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.963626 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.965734 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.965796 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.966072 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970669 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970743 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970774 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970912 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970933 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970918 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970976 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.970922 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.971051 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.971142 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.971156 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.972009 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.973297 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.973485 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.973598 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.974656 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.974878 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.984118 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 02 10:19:24 crc kubenswrapper[4679]: I1202 10:19:24.993230 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.012248 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024746 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024790 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024810 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024827 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024846 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024862 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024877 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024891 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024905 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024920 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024935 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024950 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024966 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024980 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.024995 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025009 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025024 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025041 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025058 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025075 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025091 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025105 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025122 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025145 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025134 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025165 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025196 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025221 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025242 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025263 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025285 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025336 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025345 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025359 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025380 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025400 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025420 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025439 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025463 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025484 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025504 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025507 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025524 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025567 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025598 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025622 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025643 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025664 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025688 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025711 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025734 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025753 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025772 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025794 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025813 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025833 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025854 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025875 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025896 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025917 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025939 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025961 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025981 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026003 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026023 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026045 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026067 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026093 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026116 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026142 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026163 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026186 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026209 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026242 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026266 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026290 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026327 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026350 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026372 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026395 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026419 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026442 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026484 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026517 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026541 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026566 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026588 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026611 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026635 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026658 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026682 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026704 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026740 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026880 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026907 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026931 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026956 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026980 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027021 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027046 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027070 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027096 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027122 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027148 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027173 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027197 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027222 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027247 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027350 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027544 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027785 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027847 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027907 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027935 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027958 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027995 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028020 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028044 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028068 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028090 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028111 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028134 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028156 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028182 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028209 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028237 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028260 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028282 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028320 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028339 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028361 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028383 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028407 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028431 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028451 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028474 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028496 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028518 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028542 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028561 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028580 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028599 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028620 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028670 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028703 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028733 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028761 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028788 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028817 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028849 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028877 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028906 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028936 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028966 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028996 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029028 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029057 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029085 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029115 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029145 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029174 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029200 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029233 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029267 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029320 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029354 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029384 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029408 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029439 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029470 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029499 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029529 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029562 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029590 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029635 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029670 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029699 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029726 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029759 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029785 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029814 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029843 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029873 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029903 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029927 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029948 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029971 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029992 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.030013 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.030032 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.030055 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.025662 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026561 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026703 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.026925 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027062 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027219 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027692 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027846 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027992 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.027998 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028175 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028172 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028191 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028210 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028207 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028225 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028425 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028442 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028642 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032507 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028725 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028762 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028779 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028953 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.028977 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029022 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029079 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029116 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029191 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029209 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029244 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029269 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029295 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029482 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029521 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029543 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029591 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.029612 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031118 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031390 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031581 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031593 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031717 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031765 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.031882 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032030 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032166 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032334 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032337 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032618 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032685 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032769 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032807 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032832 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033405 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033437 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033463 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033488 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033512 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033595 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt6dq\" (UniqueName: \"kubernetes.io/projected/06f77afa-f710-437b-9cfe-10959bdc1ac8-kube-api-access-bt6dq\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033628 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033652 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-socket-dir-parent\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033673 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-os-release\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033694 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/09d3af27-a3cd-4f88-a70f-941b22089a16-cni-binary-copy\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033730 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032828 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032842 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.032954 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033104 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033228 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033273 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033352 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033353 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033427 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033499 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033687 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033704 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033922 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.033998 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034034 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034140 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034183 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034298 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034434 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034437 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034469 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034524 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034563 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034781 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034805 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.034876 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.035103 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.035471 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.035701 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.035741 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.036113 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.036371 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.036648 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.036700 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.036743 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037134 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037380 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037615 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037661 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037675 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037779 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037824 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.037895 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038053 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038127 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038170 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038249 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038437 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038467 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038508 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.038794 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039026 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039066 4679 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039624 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-cni-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039670 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-os-release\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039701 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039747 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039780 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e687ed45-c025-48d3-9eeb-d4cffad70f50-hosts-file\") pod \"node-resolver-7pw64\" (UID: \"e687ed45-c025-48d3-9eeb-d4cffad70f50\") " pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039802 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039864 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039888 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-proxy-tls\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039912 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-cnibin\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039937 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039958 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040007 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06f77afa-f710-437b-9cfe-10959bdc1ac8-cni-binary-copy\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040028 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-mcd-auth-proxy-config\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040048 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-system-cni-dir\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040066 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040087 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040104 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-system-cni-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040127 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-netns\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040150 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpck8\" (UniqueName: \"kubernetes.io/projected/09d3af27-a3cd-4f88-a70f-941b22089a16-kube-api-access-wpck8\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040168 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-etc-kubernetes\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040189 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.045927 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.049921 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039130 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039394 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039507 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.039753 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040095 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040664 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.040937 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041068 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041151 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041294 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041313 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041513 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041670 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041796 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.041911 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042006 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042131 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.054690 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-k8s-cni-cncf-io\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042392 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042605 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042771 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042803 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.042846 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043156 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043210 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043321 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043384 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043420 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043570 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043667 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043691 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043692 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043782 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.043866 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044015 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044168 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044255 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044272 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044279 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055079 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044455 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.054957 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044724 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044745 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044788 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055205 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/09d3af27-a3cd-4f88-a70f-941b22089a16-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044972 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.044992 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.045122 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.045195 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.045370 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.047257 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.047346 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.047589 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.047641 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.047875 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.048159 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.048489 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.048621 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.048945 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.049268 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.049526 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.049595 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.052614 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.052621 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.052818 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.054862 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.055218 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:19:25.55519729 +0000 UTC m=+18.885336150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055526 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzzv7\" (UniqueName: \"kubernetes.io/projected/e687ed45-c025-48d3-9eeb-d4cffad70f50-kube-api-access-gzzv7\") pod \"node-resolver-7pw64\" (UID: \"e687ed45-c025-48d3-9eeb-d4cffad70f50\") " pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055563 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-multus-certs\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055598 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055628 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-cni-bin\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055652 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-rootfs\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055672 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lbf6\" (UniqueName: \"kubernetes.io/projected/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-kube-api-access-8lbf6\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055697 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-cnibin\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055721 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-hostroot\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055743 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-conf-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055761 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-daemon-config\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055783 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055804 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-cni-multus\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055824 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-kubelet\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055846 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.055912 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.055991 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:25.555979042 +0000 UTC m=+18.886117902 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.056221 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056270 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056336 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.056359 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:25.556346583 +0000 UTC m=+18.886485513 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056380 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056397 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056411 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056424 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056436 4679 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056448 4679 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056460 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056473 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056485 4679 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056497 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056509 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056524 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056538 4679 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056550 4679 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056561 4679 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056571 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056582 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056595 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056607 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056619 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056632 4679 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056644 4679 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056655 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056666 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056678 4679 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056692 4679 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056707 4679 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056722 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056735 4679 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056747 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056758 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056770 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056783 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056795 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056809 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056821 4679 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056833 4679 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056845 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056856 4679 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056867 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056879 4679 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056890 4679 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056903 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056914 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056927 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056938 4679 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056949 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056961 4679 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056971 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056983 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.056995 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057006 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057021 4679 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057032 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057043 4679 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057055 4679 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057066 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057079 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057091 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057102 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057114 4679 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057127 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057139 4679 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057150 4679 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057169 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057180 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057192 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057204 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057216 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057228 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057240 4679 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057252 4679 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057263 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057275 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057286 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057297 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057334 4679 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057345 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057397 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057413 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057425 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057438 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057451 4679 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057463 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057474 4679 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057488 4679 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057500 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057511 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057522 4679 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057534 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057546 4679 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057557 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057568 4679 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057579 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057591 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057605 4679 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057620 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057631 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057642 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057653 4679 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057665 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057650 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057677 4679 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057741 4679 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057752 4679 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057767 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057779 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057788 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057799 4679 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057808 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057817 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057828 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057837 4679 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057845 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057853 4679 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057861 4679 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057870 4679 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057883 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057892 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057901 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057909 4679 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057917 4679 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057928 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057937 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057948 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057961 4679 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057972 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057986 4679 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.057997 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058021 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058034 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058044 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058054 4679 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058065 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058073 4679 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058082 4679 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058091 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058099 4679 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058107 4679 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058116 4679 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058125 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058134 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058142 4679 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058151 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058161 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058173 4679 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058182 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058191 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058200 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058211 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058219 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058228 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058237 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058245 4679 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058254 4679 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058263 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058271 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058281 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058290 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058445 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.060687 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061041 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.058298 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061581 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061592 4679 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061614 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061624 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061632 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061641 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061654 4679 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.061664 4679 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.062006 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.062132 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.063379 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.067806 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.067936 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.067971 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.068436 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.068779 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.068784 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.068893 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.068950 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.069210 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.069629 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.069844 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.070442 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.070467 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.070472 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.072747 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.073138 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.073185 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.095699 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.162909 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-cni-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.162949 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-os-release\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.162985 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e687ed45-c025-48d3-9eeb-d4cffad70f50-hosts-file\") pod \"node-resolver-7pw64\" (UID: \"e687ed45-c025-48d3-9eeb-d4cffad70f50\") " pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163002 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163029 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-proxy-tls\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163066 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-cnibin\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163118 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163144 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06f77afa-f710-437b-9cfe-10959bdc1ac8-cni-binary-copy\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163165 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-mcd-auth-proxy-config\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163181 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-system-cni-dir\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163197 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163221 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-system-cni-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163237 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-netns\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163255 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpck8\" (UniqueName: \"kubernetes.io/projected/09d3af27-a3cd-4f88-a70f-941b22089a16-kube-api-access-wpck8\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163271 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-etc-kubernetes\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163351 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-k8s-cni-cncf-io\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163375 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/09d3af27-a3cd-4f88-a70f-941b22089a16-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163392 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzzv7\" (UniqueName: \"kubernetes.io/projected/e687ed45-c025-48d3-9eeb-d4cffad70f50-kube-api-access-gzzv7\") pod \"node-resolver-7pw64\" (UID: \"e687ed45-c025-48d3-9eeb-d4cffad70f50\") " pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163409 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-multus-certs\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163431 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-cni-bin\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163482 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-cni-bin\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163541 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-os-release\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163624 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-rootfs\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163715 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-cnibin\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163731 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lbf6\" (UniqueName: \"kubernetes.io/projected/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-kube-api-access-8lbf6\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163766 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-cnibin\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163807 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-hostroot\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163841 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-conf-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163862 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e687ed45-c025-48d3-9eeb-d4cffad70f50-hosts-file\") pod \"node-resolver-7pw64\" (UID: \"e687ed45-c025-48d3-9eeb-d4cffad70f50\") " pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163866 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-daemon-config\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163921 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-cni-multus\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163941 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-system-cni-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163950 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-kubelet\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163971 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-netns\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163982 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt6dq\" (UniqueName: \"kubernetes.io/projected/06f77afa-f710-437b-9cfe-10959bdc1ac8-kube-api-access-bt6dq\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164004 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-socket-dir-parent\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164024 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-os-release\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164047 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/09d3af27-a3cd-4f88-a70f-941b22089a16-cni-binary-copy\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164085 4679 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164099 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164113 4679 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164127 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164140 4679 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164152 4679 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164164 4679 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164176 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164188 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164200 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164211 4679 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164223 4679 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164235 4679 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164246 4679 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164257 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164271 4679 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164282 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164293 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164322 4679 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164333 4679 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164346 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164358 4679 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.163714 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-cni-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164531 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-rootfs\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164381 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-multus-certs\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164734 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-hostroot\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164781 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-cnibin\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164813 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-etc-kubernetes\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164842 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-system-cni-dir\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164870 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164897 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-run-k8s-cni-cncf-io\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164930 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-conf-dir\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164958 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-kubelet\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164998 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.164999 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-daemon-config\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.165066 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/09d3af27-a3cd-4f88-a70f-941b22089a16-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.165099 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06f77afa-f710-437b-9cfe-10959bdc1ac8-cni-binary-copy\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.165118 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-multus-socket-dir-parent\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.165077 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/06f77afa-f710-437b-9cfe-10959bdc1ac8-host-var-lib-cni-multus\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.165265 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-os-release\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.165346 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-mcd-auth-proxy-config\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.166362 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/09d3af27-a3cd-4f88-a70f-941b22089a16-cni-binary-copy\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.168447 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-proxy-tls\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.175112 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.241988 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.242448 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.245645 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/09d3af27-a3cd-4f88-a70f-941b22089a16-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.249156 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.249187 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.249206 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.249410 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:25.74925973 +0000 UTC m=+19.079398790 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.257120 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.257175 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.257193 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.257280 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:25.75725408 +0000 UTC m=+19.087392950 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.258192 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.259461 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.265115 4679 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.267845 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.271917 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.272226 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.273145 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.281925 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pjb6d"] Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.284811 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.285455 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzzv7\" (UniqueName: \"kubernetes.io/projected/e687ed45-c025-48d3-9eeb-d4cffad70f50-kube-api-access-gzzv7\") pod \"node-resolver-7pw64\" (UID: \"e687ed45-c025-48d3-9eeb-d4cffad70f50\") " pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.287492 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lbf6\" (UniqueName: \"kubernetes.io/projected/5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb-kube-api-access-8lbf6\") pod \"machine-config-daemon-lzf8q\" (UID: \"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\") " pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.291058 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.291751 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.291908 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.292102 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.294034 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.294496 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.294502 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.295057 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.297550 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7pw64" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.299120 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpck8\" (UniqueName: \"kubernetes.io/projected/09d3af27-a3cd-4f88-a70f-941b22089a16-kube-api-access-wpck8\") pod \"multus-additional-cni-plugins-2lmfz\" (UID: \"09d3af27-a3cd-4f88-a70f-941b22089a16\") " pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.299893 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt6dq\" (UniqueName: \"kubernetes.io/projected/06f77afa-f710-437b-9cfe-10959bdc1ac8-kube-api-access-bt6dq\") pod \"multus-r5sft\" (UID: \"06f77afa-f710-437b-9cfe-10959bdc1ac8\") " pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.303127 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.312646 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: W1202 10:19:25.327904 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09d3af27_a3cd_4f88_a70f_941b22089a16.slice/crio-8efe450c6510e04159e3ce5e1850bbe31ae156b94ee7525d0f80bf53474df91b WatchSource:0}: Error finding container 8efe450c6510e04159e3ce5e1850bbe31ae156b94ee7525d0f80bf53474df91b: Status 404 returned error can't find the container with id 8efe450c6510e04159e3ce5e1850bbe31ae156b94ee7525d0f80bf53474df91b Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.331615 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.342526 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.350431 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.363413 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.365936 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.374700 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.383586 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.390838 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.400254 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.415466 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.423964 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.432591 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.441125 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.450238 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.459248 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.466928 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-script-lib\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.466988 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-slash\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467012 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovn-node-metrics-cert\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467036 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467061 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-systemd-units\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467110 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmbpj\" (UniqueName: \"kubernetes.io/projected/a66c9e2c-2ca3-4348-84cc-19f365505c9e-kube-api-access-cmbpj\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467168 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-etc-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467378 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-ovn-kubernetes\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467440 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-netns\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467473 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-var-lib-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467500 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-node-log\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467584 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467634 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-systemd\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467656 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-ovn\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467678 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-config\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467726 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-log-socket\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467747 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-netd\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467808 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-bin\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467837 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-env-overrides\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467878 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-kubelet\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.467922 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.552810 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.560927 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.568914 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.569047 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:19:26.569019513 +0000 UTC m=+19.899158373 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569095 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-systemd\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569128 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-ovn\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569145 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-config\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569161 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-log-socket\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569175 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-netd\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569218 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-bin\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569232 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-env-overrides\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569248 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-kubelet\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569275 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-script-lib\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569297 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-slash\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569330 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovn-node-metrics-cert\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569348 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569366 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569384 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-systemd-units\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569398 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmbpj\" (UniqueName: \"kubernetes.io/projected/a66c9e2c-2ca3-4348-84cc-19f365505c9e-kube-api-access-cmbpj\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569414 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-etc-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569428 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-ovn-kubernetes\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569454 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-netns\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569470 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-var-lib-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569485 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-node-log\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569513 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569583 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.569739 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.569773 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:26.569766574 +0000 UTC m=+19.899905434 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569858 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-ovn\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569886 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-kubelet\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569954 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-bin\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.569946 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-etc-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570016 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-log-socket\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570028 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-slash\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570066 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-netd\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570085 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-var-lib-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.570094 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570098 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-systemd\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.570124 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:26.570114534 +0000 UTC m=+19.900253394 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570122 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-ovn-kubernetes\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570144 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-netns\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570159 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-openvswitch\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570181 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-systemd-units\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570216 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570251 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-node-log\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570587 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-env-overrides\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.570602 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-config\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.571245 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-script-lib\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.575843 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovn-node-metrics-cert\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.576004 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-r5sft" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.584142 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.586374 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmbpj\" (UniqueName: \"kubernetes.io/projected/a66c9e2c-2ca3-4348-84cc-19f365505c9e-kube-api-access-cmbpj\") pod \"ovnkube-node-pjb6d\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.617162 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:25 crc kubenswrapper[4679]: W1202 10:19:25.634348 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5df42f6a_dfd2_4c19_849c_d6ea75a1cbcb.slice/crio-0b4f715985887a3d76812286aa2a603cbd5c84102963a59c545425a4af8a4e12 WatchSource:0}: Error finding container 0b4f715985887a3d76812286aa2a603cbd5c84102963a59c545425a4af8a4e12: Status 404 returned error can't find the container with id 0b4f715985887a3d76812286aa2a603cbd5c84102963a59c545425a4af8a4e12 Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.771695 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:25 crc kubenswrapper[4679]: I1202 10:19:25.772074 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.771904 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772148 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772164 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772203 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772223 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772236 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772287 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:26.772271537 +0000 UTC m=+20.102410397 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:25 crc kubenswrapper[4679]: E1202 10:19:25.772327 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:26.772317599 +0000 UTC m=+20.102456479 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.008970 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" exitCode=0 Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.009103 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.009218 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"01cb27479623951c56ca929f70f828f97b7d6423b7afa1e9ad4da9d63a886485"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.011107 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ad8568959b495e334ed2d988b8ba43cb7b09a44490bf790fc65dfc956632090f"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.013914 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.016032 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f" exitCode=255 Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.016134 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.017934 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.017984 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.018005 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"0b4f715985887a3d76812286aa2a603cbd5c84102963a59c545425a4af8a4e12"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.019633 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.019683 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3de178f947ea9c6b82f903d113f4ad1040ea4dab9b7090decbb5536cdcc99f63"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.021626 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.021680 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.021694 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a76211315c1a04b048845d2c24b31dc7028c4c963714aeac07262626e3139e0e"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.023876 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerStarted","Data":"e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.023917 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerStarted","Data":"112db7a95389db46602bf75c9c45abe1f28d7c2f68a8a54adca75078e8ba1b4f"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.025818 4679 generic.go:334] "Generic (PLEG): container finished" podID="09d3af27-a3cd-4f88-a70f-941b22089a16" containerID="a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347" exitCode=0 Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.025875 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerDied","Data":"a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.026236 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerStarted","Data":"8efe450c6510e04159e3ce5e1850bbe31ae156b94ee7525d0f80bf53474df91b"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.027214 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.027693 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7pw64" event={"ID":"e687ed45-c025-48d3-9eeb-d4cffad70f50","Type":"ContainerStarted","Data":"fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.027872 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7pw64" event={"ID":"e687ed45-c025-48d3-9eeb-d4cffad70f50","Type":"ContainerStarted","Data":"b049b93e9d8f24ec27edf29b054ac6d31cda68ea85c782d2f1c45d0d4a5c3e14"} Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.028894 4679 scope.go:117] "RemoveContainer" containerID="9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.037492 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.053630 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.067823 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.078972 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.091016 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.105395 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.117962 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.129959 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.143351 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.156820 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.174454 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.188521 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.202490 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.214076 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.225162 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.243318 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.255085 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.267650 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.282255 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.309481 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.324328 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.339044 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.362129 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.584707 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.584850 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.584906 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:19:28.584879033 +0000 UTC m=+21.915017893 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.584970 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.584975 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.585030 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:28.585013597 +0000 UTC m=+21.915152537 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.585063 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.585116 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:28.585103739 +0000 UTC m=+21.915242599 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.742929 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5xr66"] Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.743267 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.745345 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.745539 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.745714 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.745853 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.785838 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.785889 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgcn5\" (UniqueName: \"kubernetes.io/projected/c4b4627f-c062-4e2e-a17b-cb749e9dd917-kube-api-access-kgcn5\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.785911 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.785938 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c4b4627f-c062-4e2e-a17b-cb749e9dd917-serviceca\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.785960 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4b4627f-c062-4e2e-a17b-cb749e9dd917-host\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.786889 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.786954 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.786975 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.787368 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.787422 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.787471 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.787539 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:28.787030385 +0000 UTC m=+22.117169285 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.787590 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:28.787571741 +0000 UTC m=+22.117710601 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.789691 4679 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 02 10:19:26 crc kubenswrapper[4679]: W1202 10:19:26.790105 4679 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.790435 4679 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/events/network-check-target-xd92c.187d5eb61c32a0fb\": read tcp 38.102.83.234:37448->38.102.83.234:6443: use of closed network connection" event="&Event{ObjectMeta:{network-check-target-xd92c.187d5eb61c32a0fb openshift-network-diagnostics 26713 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-network-diagnostics,Name:network-check-target-xd92c,UID:3b6479f0-333b-4a96-9adf-2099afdc2447,APIVersion:v1,ResourceVersion:25004,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-cqllr\" : [object \"openshift-network-diagnostics\"/\"kube-root-ca.crt\" not registered, object \"openshift-network-diagnostics\"/\"openshift-service-ca.crt\" not registered],Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 10:19:25 +0000 UTC,LastTimestamp:2025-12-02 10:19:26.787014125 +0000 UTC m=+20.117153015,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 10:19:26 crc kubenswrapper[4679]: W1202 10:19:26.791072 4679 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.791736 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-source-55646444c4-trplf/status\": read tcp 38.102.83.234:37448->38.102.83.234:6443: use of closed network connection" Dec 02 10:19:26 crc kubenswrapper[4679]: W1202 10:19:26.792153 4679 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Dec 02 10:19:26 crc kubenswrapper[4679]: W1202 10:19:26.792246 4679 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.857776 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.887192 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgcn5\" (UniqueName: \"kubernetes.io/projected/c4b4627f-c062-4e2e-a17b-cb749e9dd917-kube-api-access-kgcn5\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.887258 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c4b4627f-c062-4e2e-a17b-cb749e9dd917-serviceca\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.887298 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4b4627f-c062-4e2e-a17b-cb749e9dd917-host\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.887385 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4b4627f-c062-4e2e-a17b-cb749e9dd917-host\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.888900 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c4b4627f-c062-4e2e-a17b-cb749e9dd917-serviceca\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.890075 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.907902 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.910487 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.910603 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.910663 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.910709 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.910752 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:26 crc kubenswrapper[4679]: E1202 10:19:26.910796 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.912966 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.913790 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.915199 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.915978 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.917141 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.917796 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.918539 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.920935 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.921876 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.923177 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.923902 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.925257 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgcn5\" (UniqueName: \"kubernetes.io/projected/c4b4627f-c062-4e2e-a17b-cb749e9dd917-kube-api-access-kgcn5\") pod \"node-ca-5xr66\" (UID: \"c4b4627f-c062-4e2e-a17b-cb749e9dd917\") " pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.925329 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.926002 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.926884 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.927867 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.928483 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.928692 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.929849 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.930398 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.931099 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.932238 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.932768 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.933976 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.934509 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.935959 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.936754 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.937491 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.938767 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.940426 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.943261 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.943860 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.944809 4679 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.944925 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.945662 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.946868 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.948063 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.949107 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.950586 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.951277 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.952350 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.953055 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.954099 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.954634 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.955935 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.956764 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.957762 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.958274 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.959254 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.959899 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.961098 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.961719 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.962615 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.962658 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.963258 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.964401 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.964999 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.965466 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 02 10:19:26 crc kubenswrapper[4679]: I1202 10:19:26.982742 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.003159 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.017051 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.029676 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.032335 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.032378 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.032387 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.034217 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.035822 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.036340 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.037506 4679 generic.go:334] "Generic (PLEG): container finished" podID="09d3af27-a3cd-4f88-a70f-941b22089a16" containerID="2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c" exitCode=0 Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.037599 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerDied","Data":"2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.045981 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.072650 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.087001 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.104044 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.114857 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.129587 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.144410 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.160882 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.175520 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.197386 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5xr66" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.209783 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: W1202 10:19:27.218282 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4b4627f_c062_4e2e_a17b_cb749e9dd917.slice/crio-06122d93d5643d70874bdd32364cd39b3d7a677a04f1b5f87318e8a71e65c83d WatchSource:0}: Error finding container 06122d93d5643d70874bdd32364cd39b3d7a677a04f1b5f87318e8a71e65c83d: Status 404 returned error can't find the container with id 06122d93d5643d70874bdd32364cd39b3d7a677a04f1b5f87318e8a71e65c83d Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.237051 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.251663 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.266429 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.284534 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.298997 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.313152 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.325287 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.340201 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.353415 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.370558 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.380449 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.395637 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.420356 4679 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.422095 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.422130 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.422139 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.422230 4679 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.452671 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.458226 4679 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.458521 4679 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.459645 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.459693 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.459704 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.459723 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.459750 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.466746 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: E1202 10:19:27.477876 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.480616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.480657 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.480665 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.480680 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.480690 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.487609 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: E1202 10:19:27.493600 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.501957 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.502004 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.502017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.502033 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.502045 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.505400 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: E1202 10:19:27.513756 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.517178 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.517234 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.517245 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.517260 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.517271 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.520406 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: E1202 10:19:27.530909 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.534051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.534083 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.534113 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.534130 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.534140 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.534082 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: E1202 10:19:27.545622 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:27 crc kubenswrapper[4679]: E1202 10:19:27.545737 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.547442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.547488 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.547499 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.547543 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.547554 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.650278 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.650340 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.650352 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.650372 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.650383 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.742344 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.752729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.752762 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.752772 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.752787 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.752798 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.855321 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.855369 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.855381 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.855398 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.855411 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.951617 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.958031 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.958067 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.958079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.958095 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:27 crc kubenswrapper[4679]: I1202 10:19:27.958107 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:27Z","lastTransitionTime":"2025-12-02T10:19:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.043086 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5xr66" event={"ID":"c4b4627f-c062-4e2e-a17b-cb749e9dd917","Type":"ContainerStarted","Data":"0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.043138 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5xr66" event={"ID":"c4b4627f-c062-4e2e-a17b-cb749e9dd917","Type":"ContainerStarted","Data":"06122d93d5643d70874bdd32364cd39b3d7a677a04f1b5f87318e8a71e65c83d"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.044706 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.045874 4679 generic.go:334] "Generic (PLEG): container finished" podID="09d3af27-a3cd-4f88-a70f-941b22089a16" containerID="a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1" exitCode=0 Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.045939 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerDied","Data":"a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.050751 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.050803 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.050820 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.063188 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.063227 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.063238 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.063253 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.063264 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.070609 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.099738 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.111703 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.124203 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.137539 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.150393 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.161776 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.165356 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.165399 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.165415 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.165434 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.165445 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.174210 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.188646 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.199015 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.210656 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.222442 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.232219 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.246838 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.257254 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.267132 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.267170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.267179 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.267192 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.267206 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.268656 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.279038 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.298627 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.309427 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.315115 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.321603 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.333611 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.344040 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.357806 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.369803 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.369845 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.369857 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.369873 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.369884 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.372859 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.386294 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.409321 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.471995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.472033 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.472041 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.472055 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.472064 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.574243 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.574279 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.574288 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.574317 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.574335 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.601876 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.601989 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.602036 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.602095 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:19:32.602064404 +0000 UTC m=+25.932203264 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.602112 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.602173 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:32.602156977 +0000 UTC m=+25.932295927 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.602181 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.602234 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:32.602223459 +0000 UTC m=+25.932362419 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.676160 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.676194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.676204 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.676220 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.676232 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.778139 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.778170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.778178 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.778192 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.778200 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.791190 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.794543 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.801052 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.803902 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.803956 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804107 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804139 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804154 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804221 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:32.804201756 +0000 UTC m=+26.134340646 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804399 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804474 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804533 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.804617 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:32.804608488 +0000 UTC m=+26.134747338 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.807248 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.820292 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.830190 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.842558 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.854533 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.866610 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.878687 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.880153 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.880187 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.880197 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.880213 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.880223 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.890961 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.904868 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.908355 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.908461 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.908361 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.908528 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.908354 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:28 crc kubenswrapper[4679]: E1202 10:19:28.908578 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.921139 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.933618 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.947056 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.960871 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.982802 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.982844 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.982855 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.982874 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.982885 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:28Z","lastTransitionTime":"2025-12-02T10:19:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:28 crc kubenswrapper[4679]: I1202 10:19:28.987731 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.026746 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.055821 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.058277 4679 generic.go:334] "Generic (PLEG): container finished" podID="09d3af27-a3cd-4f88-a70f-941b22089a16" containerID="fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad" exitCode=0 Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.058763 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerDied","Data":"fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.069314 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.084409 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.084434 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.084442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.084454 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.084463 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.106539 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.147346 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.186919 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.187205 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.187224 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.187233 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.187245 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.187254 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.228415 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.267734 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.289791 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.289827 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.289837 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.289852 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.289863 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.308893 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.351522 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.394169 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.394216 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.394230 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.394248 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.394261 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.394395 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.429812 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.468786 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.496037 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.496076 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.496086 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.496101 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.496112 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.507884 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.547349 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.589287 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.597891 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.597945 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.597956 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.597973 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.597983 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.629653 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.665975 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.700507 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.700542 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.700552 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.700601 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.700614 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.707371 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.753001 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.791424 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.802864 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.802889 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.802898 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.802912 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.802921 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.827780 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.888266 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.905179 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.905204 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.905212 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.905223 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.905232 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:29Z","lastTransitionTime":"2025-12-02T10:19:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.910051 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.948254 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:29 crc kubenswrapper[4679]: I1202 10:19:29.987861 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.006935 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.006982 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.006996 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.007017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.007034 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.031327 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.064982 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.068657 4679 generic.go:334] "Generic (PLEG): container finished" podID="09d3af27-a3cd-4f88-a70f-941b22089a16" containerID="82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7" exitCode=0 Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.068715 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerDied","Data":"82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.068754 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.108462 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.108494 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.108502 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.108514 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.108523 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.113467 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.147981 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.189704 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.211097 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.211143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.211154 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.211170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.211188 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.230370 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.268430 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.307937 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.313698 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.313732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.313740 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.313753 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.313782 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.346765 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.386934 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.416868 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.416907 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.416917 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.416931 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.416941 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.426015 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.467558 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.507013 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.518801 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.518845 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.518855 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.518870 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.518881 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.548963 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.590970 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.621690 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.621731 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.621744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.621760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.621770 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.632575 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:30Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.724730 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.724771 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.724781 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.724795 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.724806 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.827634 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.827684 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.827700 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.827721 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.827738 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.908374 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.908424 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.908515 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:30 crc kubenswrapper[4679]: E1202 10:19:30.908811 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:30 crc kubenswrapper[4679]: E1202 10:19:30.909270 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:30 crc kubenswrapper[4679]: E1202 10:19:30.909353 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.930129 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.930162 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.930172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.930187 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:30 crc kubenswrapper[4679]: I1202 10:19:30.930200 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:30Z","lastTransitionTime":"2025-12-02T10:19:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.033604 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.033647 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.033661 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.033679 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.033691 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.076505 4679 generic.go:334] "Generic (PLEG): container finished" podID="09d3af27-a3cd-4f88-a70f-941b22089a16" containerID="91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b" exitCode=0 Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.076550 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerDied","Data":"91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.089060 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.101862 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.119833 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.133220 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.135727 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.135766 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.135777 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.135792 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.135803 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.144402 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.157828 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.173468 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.193024 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.205863 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.220869 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.232474 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.238091 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.238118 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.238126 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.238137 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.238147 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.242992 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.255079 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.268850 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:31Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.340637 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.340677 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.340686 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.340702 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.340713 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.443598 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.443636 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.443645 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.443659 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.443668 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.546539 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.546591 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.546614 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.546631 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.546643 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.649162 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.649207 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.649217 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.649230 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.649242 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.751981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.752025 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.752036 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.752051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.752060 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.854655 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.854705 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.854727 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.854755 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.854771 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.957932 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.957966 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.957976 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.957995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:31 crc kubenswrapper[4679]: I1202 10:19:31.958005 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:31Z","lastTransitionTime":"2025-12-02T10:19:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.061694 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.061760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.061774 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.061800 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.061815 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.164065 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.164113 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.164124 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.164142 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.164154 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.267020 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.267050 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.267059 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.267073 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.267083 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.369228 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.369272 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.369282 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.369300 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.369326 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.471982 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.472049 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.472060 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.472078 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.472091 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.574407 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.574458 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.574469 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.574485 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.574494 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.644360 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.644448 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.644476 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:19:40.644454577 +0000 UTC m=+33.974593447 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.644538 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.644596 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.644666 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.644692 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:40.644672114 +0000 UTC m=+33.974810974 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.644715 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:40.644705845 +0000 UTC m=+33.974844815 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.676507 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.676541 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.676549 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.676562 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.676572 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.779026 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.779057 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.779065 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.779077 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.779085 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.845851 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.845922 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846052 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846083 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846098 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846144 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:40.846129275 +0000 UTC m=+34.176268135 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846052 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846195 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846205 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.846448 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:40.846433604 +0000 UTC m=+34.176572464 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.881294 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.881408 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.881425 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.881450 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.881474 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.907926 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.908023 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.908082 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.907943 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.908192 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:32 crc kubenswrapper[4679]: E1202 10:19:32.908288 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.984719 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.984774 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.984789 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.984811 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:32 crc kubenswrapper[4679]: I1202 10:19:32.984829 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:32Z","lastTransitionTime":"2025-12-02T10:19:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.086092 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.086116 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.086125 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.086137 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.086145 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.088186 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" event={"ID":"09d3af27-a3cd-4f88-a70f-941b22089a16","Type":"ContainerStarted","Data":"2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.092151 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.102498 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.117519 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.130452 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.144110 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.159818 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.182292 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.188140 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.188195 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.188211 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.188229 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.188241 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.196967 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.211471 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.223903 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.237231 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.250165 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.263823 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.278021 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.290434 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.290481 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.290491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.290505 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.290517 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.291009 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.303043 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.316687 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.330474 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.341235 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.351888 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.361693 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.376355 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.390811 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.393439 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.393482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.393492 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.393510 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.393522 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.404494 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.421393 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.440414 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.453425 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.468807 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.480788 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:33Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.495378 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.495420 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.495429 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.495442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.495453 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.598578 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.598630 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.598642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.598659 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.598672 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.700948 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.701004 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.701015 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.701032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.701044 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.803281 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.803338 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.803349 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.803364 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.803375 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.905916 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.905955 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.905966 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.905981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:33 crc kubenswrapper[4679]: I1202 10:19:33.905994 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:33Z","lastTransitionTime":"2025-12-02T10:19:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.008230 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.008269 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.008279 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.008294 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.008326 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.095125 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.096088 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.096135 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.110221 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.110256 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.110264 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.110279 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.110288 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.123956 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.125296 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.134339 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.146177 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.160138 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.175035 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.187520 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.205640 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.212838 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.212888 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.212900 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.212917 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.212928 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.230294 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.245669 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.257279 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.270125 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.282665 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.292970 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.303541 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.315150 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.315200 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.315211 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.315226 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.315235 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.316025 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.327973 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.340244 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.362386 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.376601 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.386798 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.395943 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.404827 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.415422 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.416889 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.416909 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.416918 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.416930 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.416939 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.427457 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.450336 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.462157 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.483637 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.496505 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.508982 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:34Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.519698 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.519741 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.519751 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.519766 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.519778 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.622367 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.622410 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.622421 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.622437 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.622448 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.724549 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.724604 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.724612 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.724627 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.724636 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.827006 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.827056 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.827067 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.827087 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.827099 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.908735 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.908863 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:34 crc kubenswrapper[4679]: E1202 10:19:34.909086 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.909124 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:34 crc kubenswrapper[4679]: E1202 10:19:34.909198 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:34 crc kubenswrapper[4679]: E1202 10:19:34.909369 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.929578 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.929647 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.929670 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.929700 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:34 crc kubenswrapper[4679]: I1202 10:19:34.929723 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:34Z","lastTransitionTime":"2025-12-02T10:19:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.032608 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.032668 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.032684 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.032704 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.033088 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.099919 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/0.log" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.102550 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345" exitCode=1 Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.102589 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.103265 4679 scope.go:117] "RemoveContainer" containerID="9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.116809 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.131221 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.135866 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.135911 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.135926 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.135946 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.135959 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.145838 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.157783 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.168086 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.178676 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.190630 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.202394 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.215719 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.228172 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.238550 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.238594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.238604 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.238622 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.238632 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.244131 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.257061 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.268254 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.280986 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:35Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.341549 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.341591 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.341603 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.341620 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.341631 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.443522 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.443568 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.443581 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.443597 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.443609 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.545904 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.545956 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.545965 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.545980 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.545990 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.647849 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.647881 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.647891 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.647905 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.647917 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.750148 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.750214 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.750224 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.750241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.750253 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.852733 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.852769 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.852779 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.852796 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.852812 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.955163 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.955198 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.955209 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.955225 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:35 crc kubenswrapper[4679]: I1202 10:19:35.955234 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:35Z","lastTransitionTime":"2025-12-02T10:19:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.057156 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.057188 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.057200 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.057216 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.057229 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.109115 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/0.log" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.111717 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.111866 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.126292 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.140824 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.154109 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.159377 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.159412 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.159426 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.159441 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.159450 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.166823 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.185647 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.205557 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.220180 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.233580 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.245634 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.259649 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.261558 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.261601 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.261616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.261631 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.261642 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.273253 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.285088 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.295407 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.305638 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.364204 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.364256 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.364265 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.364279 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.364289 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.467038 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.467079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.467091 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.467104 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.467114 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.569502 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.569845 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.569859 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.569876 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.569887 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.672632 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.672684 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.672693 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.672708 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.672718 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.776093 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.776143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.776152 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.776174 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.776189 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.879197 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.879263 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.879286 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.879349 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.879377 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.908524 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.908587 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.908673 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:36 crc kubenswrapper[4679]: E1202 10:19:36.908759 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:36 crc kubenswrapper[4679]: E1202 10:19:36.909584 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:36 crc kubenswrapper[4679]: E1202 10:19:36.909839 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.929318 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.944575 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.961339 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.974357 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.982241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.982271 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.982279 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.982292 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.982325 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:36Z","lastTransitionTime":"2025-12-02T10:19:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:36 crc kubenswrapper[4679]: I1202 10:19:36.988942 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.009934 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.028237 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.040433 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.055688 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.076042 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.083987 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.084014 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.084023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.084035 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.084044 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.095904 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.110766 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.114907 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.123861 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.136668 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.187259 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.187323 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.187334 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.187350 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.187361 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.289744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.289790 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.289802 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.289819 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.289830 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.392683 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.392730 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.392750 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.392764 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.392783 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.495182 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.495223 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.495234 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.495248 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.495258 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.597088 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.597132 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.597143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.597158 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.597167 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.620463 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q"] Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.620971 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.622734 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.623858 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.634883 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.649587 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.661472 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.672632 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.682024 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.682066 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.682078 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.682094 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.682106 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.682909 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: E1202 10:19:37.693759 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.694332 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/141dee09-85e8-43f5-a5d5-52458cce375a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.694374 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z58h7\" (UniqueName: \"kubernetes.io/projected/141dee09-85e8-43f5-a5d5-52458cce375a-kube-api-access-z58h7\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.694408 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/141dee09-85e8-43f5-a5d5-52458cce375a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.694444 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/141dee09-85e8-43f5-a5d5-52458cce375a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.695478 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.697915 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.697966 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.697981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.698001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.698015 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.708685 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: E1202 10:19:37.710031 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.713838 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.713883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.713896 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.713917 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.713929 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.721560 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: E1202 10:19:37.725518 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.728459 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.728496 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.728535 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.728551 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.728559 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.733913 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: E1202 10:19:37.740143 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.743314 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.743459 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.743562 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.743716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.743797 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.749002 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: E1202 10:19:37.759656 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: E1202 10:19:37.759772 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.761537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.761589 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.761602 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.761618 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.761634 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.770089 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.789617 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.795162 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/141dee09-85e8-43f5-a5d5-52458cce375a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.795227 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/141dee09-85e8-43f5-a5d5-52458cce375a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.795286 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/141dee09-85e8-43f5-a5d5-52458cce375a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.795339 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z58h7\" (UniqueName: \"kubernetes.io/projected/141dee09-85e8-43f5-a5d5-52458cce375a-kube-api-access-z58h7\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.796128 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/141dee09-85e8-43f5-a5d5-52458cce375a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.796290 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/141dee09-85e8-43f5-a5d5-52458cce375a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.801048 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/141dee09-85e8-43f5-a5d5-52458cce375a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.803661 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.811658 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z58h7\" (UniqueName: \"kubernetes.io/projected/141dee09-85e8-43f5-a5d5-52458cce375a-kube-api-access-z58h7\") pod \"ovnkube-control-plane-749d76644c-vs22q\" (UID: \"141dee09-85e8-43f5-a5d5-52458cce375a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.817023 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.830800 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.863997 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.864058 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.864070 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.864088 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.864101 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.933763 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.966833 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.966904 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.966916 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.966932 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:37 crc kubenswrapper[4679]: I1202 10:19:37.966942 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:37Z","lastTransitionTime":"2025-12-02T10:19:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.069369 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.069410 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.069422 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.069453 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.069469 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.119243 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" event={"ID":"141dee09-85e8-43f5-a5d5-52458cce375a","Type":"ContainerStarted","Data":"2b3ca9d6c72767602b6b94df4cdaadcda3a1f8218f14320fa685e4ba99606c32"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.121436 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/1.log" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.122001 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/0.log" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.124175 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da" exitCode=1 Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.124223 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.124267 4679 scope.go:117] "RemoveContainer" containerID="9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.124940 4679 scope.go:117] "RemoveContainer" containerID="b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da" Dec 02 10:19:38 crc kubenswrapper[4679]: E1202 10:19:38.125084 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.136894 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.146091 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.154148 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.163548 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.172760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.172815 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.172828 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.172848 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.172863 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.174024 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.185576 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.197440 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.210120 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.222730 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.235900 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.253912 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.264945 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.277838 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.278297 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.278363 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.278375 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.278393 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.278404 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.289379 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.301537 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:38Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.381370 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.381411 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.381426 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.381444 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.381456 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.484474 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.484537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.484561 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.484641 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.484667 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.587231 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.587284 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.587299 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.587344 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.587360 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.689844 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.689884 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.689893 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.689908 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.689919 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.792867 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.792912 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.792922 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.792938 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.792950 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.894679 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.894718 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.894729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.894746 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.894759 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.908541 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.908541 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.908706 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:38 crc kubenswrapper[4679]: E1202 10:19:38.908824 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:38 crc kubenswrapper[4679]: E1202 10:19:38.908947 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:38 crc kubenswrapper[4679]: E1202 10:19:38.909111 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.997795 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.997859 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.997876 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.997903 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:38 crc kubenswrapper[4679]: I1202 10:19:38.997924 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:38Z","lastTransitionTime":"2025-12-02T10:19:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.098651 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xlg8w"] Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.099090 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: E1202 10:19:39.099149 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.099826 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.099850 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.099859 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.099870 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.099880 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.115522 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.131499 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" event={"ID":"141dee09-85e8-43f5-a5d5-52458cce375a","Type":"ContainerStarted","Data":"1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.131563 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" event={"ID":"141dee09-85e8-43f5-a5d5-52458cce375a","Type":"ContainerStarted","Data":"e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.133972 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/1.log" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.138012 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.153565 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.168212 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.183477 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.201745 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.201788 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.201797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.201812 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.201821 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.207017 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.209374 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j2jc\" (UniqueName: \"kubernetes.io/projected/c1827954-0852-40f4-ad94-c979f72addb9-kube-api-access-7j2jc\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.209509 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.224322 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.238628 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.255823 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.270446 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.287527 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.301939 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.304165 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.304232 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.304255 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.304286 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.304342 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.310508 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j2jc\" (UniqueName: \"kubernetes.io/projected/c1827954-0852-40f4-ad94-c979f72addb9-kube-api-access-7j2jc\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.310575 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: E1202 10:19:39.310717 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:39 crc kubenswrapper[4679]: E1202 10:19:39.310771 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:39.810755816 +0000 UTC m=+33.140894696 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.313517 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.323994 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.327346 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j2jc\" (UniqueName: \"kubernetes.io/projected/c1827954-0852-40f4-ad94-c979f72addb9-kube-api-access-7j2jc\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.334511 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.345169 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.356255 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.366249 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.377795 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.387789 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.397151 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.406627 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.406652 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.406660 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.406672 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.406681 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.410536 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.420135 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.436853 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.447516 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.461495 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.473528 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.488076 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.500645 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.508675 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.508747 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.508756 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.508770 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.508780 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.514256 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.526480 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.539925 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:39Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.611017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.611044 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.611053 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.611067 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.611076 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.714215 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.714287 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.714342 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.714357 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.714368 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.813724 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:39 crc kubenswrapper[4679]: E1202 10:19:39.813845 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:39 crc kubenswrapper[4679]: E1202 10:19:39.813917 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:40.813900793 +0000 UTC m=+34.144039653 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.820939 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.821009 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.821032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.821059 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.821080 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.923429 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.923469 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.923480 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.923495 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:39 crc kubenswrapper[4679]: I1202 10:19:39.923507 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:39Z","lastTransitionTime":"2025-12-02T10:19:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.025640 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.025687 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.025703 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.025727 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.025744 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.128491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.128529 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.128537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.128552 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.128561 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.230774 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.230810 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.230821 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.230836 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.230845 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.333581 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.333653 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.333673 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.333701 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.333722 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.435901 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.435950 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.435963 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.435978 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.435989 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.538538 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.538592 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.538603 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.538620 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.538631 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.641477 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.641529 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.641544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.641564 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.641580 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.722220 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.722377 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.722404 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:19:56.722370094 +0000 UTC m=+50.052508954 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.722477 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.722494 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.722525 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:56.722512438 +0000 UTC m=+50.052651298 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.722588 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.722629 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:56.722622301 +0000 UTC m=+50.052761161 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.744072 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.744110 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.744120 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.744134 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.744143 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.823135 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.823271 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.823366 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:42.823351627 +0000 UTC m=+36.153490487 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.846700 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.846744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.846756 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.846771 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.846783 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.907785 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.907831 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.907963 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.908025 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.908021 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.908168 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.908290 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.908445 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.923614 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.923672 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923769 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923769 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923786 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923791 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923795 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923801 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923835 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:56.923823276 +0000 UTC m=+50.253962136 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:40 crc kubenswrapper[4679]: E1202 10:19:40.923849 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:56.923843347 +0000 UTC m=+50.253982207 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.949594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.949671 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.949682 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.949708 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:40 crc kubenswrapper[4679]: I1202 10:19:40.949721 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:40Z","lastTransitionTime":"2025-12-02T10:19:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.052300 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.052390 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.052409 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.052432 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.052452 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.154948 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.155017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.155040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.155168 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.155210 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.258047 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.258106 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.258123 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.258146 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.258163 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.360928 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.361002 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.361022 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.361044 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.361063 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.463765 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.464079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.464110 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.464133 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.464149 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.567109 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.567164 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.567186 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.567207 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.567221 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.669992 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.670039 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.670049 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.670065 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.670075 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.772272 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.772300 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.772325 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.772340 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.772352 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.874977 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.875264 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.875368 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.875458 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.875540 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.979454 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.979526 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.979543 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.979567 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:41 crc kubenswrapper[4679]: I1202 10:19:41.979583 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:41Z","lastTransitionTime":"2025-12-02T10:19:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.068526 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.082896 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.082960 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.082976 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.082999 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.083019 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.091050 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.111225 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.134136 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.165553 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.184708 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.186594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.186663 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.186696 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.186729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.186753 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.198491 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.210393 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.224367 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.241808 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.252985 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.264432 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.273608 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.283239 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.288363 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.288402 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.288421 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.288439 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.288449 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.291966 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.304438 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.317159 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:42Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.390238 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.390289 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.390298 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.390327 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.390335 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.492561 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.492604 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.492616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.492632 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.492643 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.595087 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.595135 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.595149 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.595170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.595184 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.697688 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.697748 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.697762 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.697784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.697800 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.800029 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.800079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.800087 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.800104 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.800116 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.844925 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:42 crc kubenswrapper[4679]: E1202 10:19:42.845145 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:42 crc kubenswrapper[4679]: E1202 10:19:42.845661 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:46.845646936 +0000 UTC m=+40.175785796 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.902657 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.902695 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.902721 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.902739 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.902750 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:42Z","lastTransitionTime":"2025-12-02T10:19:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.908184 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.908248 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:42 crc kubenswrapper[4679]: E1202 10:19:42.908299 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.908363 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:42 crc kubenswrapper[4679]: E1202 10:19:42.908399 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:42 crc kubenswrapper[4679]: E1202 10:19:42.908495 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:42 crc kubenswrapper[4679]: I1202 10:19:42.908521 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:42 crc kubenswrapper[4679]: E1202 10:19:42.908585 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.005390 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.005443 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.005460 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.005519 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.005536 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.107952 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.107995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.108003 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.108021 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.108030 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.211402 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.211446 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.211457 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.211475 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.211488 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.314228 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.314264 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.314274 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.314290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.314302 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.416083 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.416123 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.416131 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.416144 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.416155 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.518646 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.518723 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.518743 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.518771 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.518793 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.621457 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.621531 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.621556 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.621592 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.621613 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.723951 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.724002 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.724017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.724037 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.724051 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.826505 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.826590 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.826612 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.826639 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.826659 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.929273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.929339 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.929351 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.929368 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:43 crc kubenswrapper[4679]: I1202 10:19:43.929381 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:43Z","lastTransitionTime":"2025-12-02T10:19:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.032011 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.032058 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.032072 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.032092 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.032107 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.134975 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.135025 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.135037 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.135057 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.135068 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.238621 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.238695 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.238710 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.238737 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.238752 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.355563 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.355598 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.355609 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.355622 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.355632 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.458389 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.458471 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.458482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.458497 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.458507 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.560797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.560862 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.560882 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.560904 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.560920 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.662800 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.662842 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.662851 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.662867 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.662876 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.766465 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.766500 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.766509 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.766528 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.766538 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.869959 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.870040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.870072 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.870099 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.870119 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.907972 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.908010 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.908011 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.908063 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:44 crc kubenswrapper[4679]: E1202 10:19:44.908204 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:44 crc kubenswrapper[4679]: E1202 10:19:44.908324 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:44 crc kubenswrapper[4679]: E1202 10:19:44.908466 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:44 crc kubenswrapper[4679]: E1202 10:19:44.908526 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.973076 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.973126 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.973162 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.973181 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:44 crc kubenswrapper[4679]: I1202 10:19:44.973198 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:44Z","lastTransitionTime":"2025-12-02T10:19:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.075554 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.075613 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.075625 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.075644 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.075659 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.178717 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.178802 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.178826 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.178857 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.178883 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.280940 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.281012 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.281024 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.281042 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.281055 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.383064 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.383120 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.383128 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.383149 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.383158 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.485326 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.485373 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.485385 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.485403 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.485414 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.587654 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.587691 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.587703 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.587717 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.587729 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.689875 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.689921 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.689930 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.689943 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.689951 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.794765 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.794855 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.794872 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.794898 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.794915 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.897699 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.897751 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.897761 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.897778 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.897790 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.999493 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.999538 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.999548 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.999562 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:45 crc kubenswrapper[4679]: I1202 10:19:45.999574 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:45Z","lastTransitionTime":"2025-12-02T10:19:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.101480 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.101517 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.101533 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.101551 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.101561 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.204011 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.204059 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.204070 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.204088 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.204098 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.305838 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.305883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.305894 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.305910 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.305921 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.409387 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.409457 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.409471 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.409487 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.409498 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.511765 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.511819 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.511833 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.511854 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.511868 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.614379 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.614420 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.614432 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.614455 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.614480 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.717150 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.717240 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.717264 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.717299 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.717369 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.819509 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.819538 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.819545 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.819557 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.819565 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.892550 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:46 crc kubenswrapper[4679]: E1202 10:19:46.892697 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:46 crc kubenswrapper[4679]: E1202 10:19:46.892749 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:19:54.892735164 +0000 UTC m=+48.222874014 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.908164 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.908215 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.908215 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:46 crc kubenswrapper[4679]: E1202 10:19:46.908288 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:46 crc kubenswrapper[4679]: E1202 10:19:46.908349 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:46 crc kubenswrapper[4679]: E1202 10:19:46.908394 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.908477 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:46 crc kubenswrapper[4679]: E1202 10:19:46.908747 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.921699 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.921746 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.921758 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.921775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.921786 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:46Z","lastTransitionTime":"2025-12-02T10:19:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.925642 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:46Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.937401 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:46Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.951261 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:46Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.964779 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:46Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.982391 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b4f0373df4048f4fba2d6be8e6372414f9c585eae9d8d831929a306261e3345\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:34Z\\\",\\\"message\\\":\\\"opping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249084 5953 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.248859 5953 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249166 5953 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249325 5953 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249450 5953 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 10:19:34.249701 5953 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 10:19:34.249983 5953 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 10:19:34.250045 5953 factory.go:656] Stopping watch factory\\\\nI1202 10:19:34.250061 5953 ovnkube.go:599] Stopped ovnkube\\\\nI1202 10:19:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:46Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:46 crc kubenswrapper[4679]: I1202 10:19:46.993655 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:46Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.005829 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.022905 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.023343 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.023378 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.023388 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.023403 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.023414 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.034672 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.045662 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.056135 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.066288 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.075002 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.085738 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.094654 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.105453 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.124943 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.124986 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.124996 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.125010 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.125021 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.226979 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.227027 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.227042 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.227064 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.227080 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.329866 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.329918 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.329929 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.329947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.329963 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.422353 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.423245 4679 scope.go:117] "RemoveContainer" containerID="b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da" Dec 02 10:19:47 crc kubenswrapper[4679]: E1202 10:19:47.423406 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.432487 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.432524 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.432535 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.432550 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.432562 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.439442 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.452207 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.467642 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.479542 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.490971 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.505367 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.515898 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.527686 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.534145 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.534180 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.534190 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.534205 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.534215 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.539011 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.551365 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.564788 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.582475 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.600755 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.612289 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.623508 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.636492 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.636559 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.636582 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.636613 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.636636 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.637020 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.739791 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.739838 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.739850 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.739866 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.739877 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.842918 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.842978 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.842995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.843019 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.843034 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.905385 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.905453 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.905476 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.905537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.905562 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: E1202 10:19:47.947075 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.951058 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.951105 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.951115 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.951131 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.951141 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: E1202 10:19:47.971151 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.974781 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.975102 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.975114 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.975128 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.975139 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:47 crc kubenswrapper[4679]: E1202 10:19:47.987797 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:47Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.991533 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.991558 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.991565 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.991579 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:47 crc kubenswrapper[4679]: I1202 10:19:47.991587 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:47Z","lastTransitionTime":"2025-12-02T10:19:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.002967 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:48Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.005656 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.005685 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.005693 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.005703 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.005713 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.016496 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:48Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.016669 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.018157 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.018188 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.018198 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.018213 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.018223 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.121064 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.121125 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.121136 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.121150 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.121161 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.222666 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.222708 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.222720 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.222735 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.222745 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.325248 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.325358 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.325375 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.325398 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.325416 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.428760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.428829 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.428853 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.428884 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.428905 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.531679 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.531747 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.531759 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.531782 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.531794 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.634635 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.634716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.634744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.634775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.634800 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.737346 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.737388 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.737397 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.737412 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.737423 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.840111 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.840154 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.840165 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.840182 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.840196 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.908195 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.908267 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.908392 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.908480 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.908593 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.908714 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.908906 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:48 crc kubenswrapper[4679]: E1202 10:19:48.909001 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.943217 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.943289 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.943322 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.943340 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:48 crc kubenswrapper[4679]: I1202 10:19:48.943353 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:48Z","lastTransitionTime":"2025-12-02T10:19:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.047524 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.047598 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.047616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.047642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.047661 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.150732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.150803 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.150823 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.150850 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.150873 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.253633 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.253696 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.253713 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.253732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.253749 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.356564 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.356636 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.356656 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.356681 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.356702 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.460238 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.460328 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.460345 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.460370 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.460391 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.563328 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.563371 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.563380 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.563395 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.563407 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.666366 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.666429 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.666447 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.666472 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.666491 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.769285 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.769343 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.769354 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.769369 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.769379 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.871225 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.871272 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.871281 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.871301 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.871333 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.972937 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.972973 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.972992 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.973008 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:49 crc kubenswrapper[4679]: I1202 10:19:49.973019 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:49Z","lastTransitionTime":"2025-12-02T10:19:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.076025 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.076071 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.076080 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.076095 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.076104 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.179203 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.179267 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.179288 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.179341 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.179361 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.281778 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.281807 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.281818 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.281831 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.281840 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.383884 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.383930 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.383945 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.383959 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.383970 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.486714 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.486744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.486752 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.486765 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.486773 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.588647 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.588690 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.588699 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.588726 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.588740 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.690679 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.690736 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.690755 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.690775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.690789 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.793120 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.793165 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.793173 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.793187 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.793196 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.895241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.895285 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.895295 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.895323 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.895333 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.908017 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.908076 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:50 crc kubenswrapper[4679]: E1202 10:19:50.908127 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:50 crc kubenswrapper[4679]: E1202 10:19:50.908193 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.908241 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.908329 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:50 crc kubenswrapper[4679]: E1202 10:19:50.908446 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:50 crc kubenswrapper[4679]: E1202 10:19:50.908566 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.997591 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.997680 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.997697 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.997748 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:50 crc kubenswrapper[4679]: I1202 10:19:50.997768 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:50Z","lastTransitionTime":"2025-12-02T10:19:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.101186 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.101229 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.101239 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.101257 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.101272 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.204371 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.204422 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.204433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.204450 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.204460 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.307237 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.307299 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.307355 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.307381 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.307398 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.409894 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.409962 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.409977 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.409993 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.410004 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.512380 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.512419 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.512428 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.512442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.512452 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.613960 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.614013 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.614024 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.614040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.614051 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.716024 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.716076 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.716093 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.716118 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.716138 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.818814 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.818868 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.818884 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.818906 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.818924 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.921766 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.921806 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.921815 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.921834 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:51 crc kubenswrapper[4679]: I1202 10:19:51.921845 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:51Z","lastTransitionTime":"2025-12-02T10:19:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.024816 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.024883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.024907 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.024940 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.024962 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.128123 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.128179 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.128194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.128215 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.128231 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.231356 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.231397 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.231406 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.231424 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.231436 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.333086 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.333137 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.333151 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.333167 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.333176 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.435479 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.435511 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.435520 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.435534 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.435542 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.564449 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.564507 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.564522 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.564544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.564560 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.667289 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.667380 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.667391 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.667408 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.667422 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.769102 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.769146 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.769159 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.769174 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.769182 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.872730 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.872775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.872784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.872800 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.872814 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.908505 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.908588 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.908544 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:52 crc kubenswrapper[4679]: E1202 10:19:52.908765 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.908798 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:52 crc kubenswrapper[4679]: E1202 10:19:52.908962 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:52 crc kubenswrapper[4679]: E1202 10:19:52.909091 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:52 crc kubenswrapper[4679]: E1202 10:19:52.909399 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.975422 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.975531 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.975540 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.975558 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:52 crc kubenswrapper[4679]: I1202 10:19:52.975567 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:52Z","lastTransitionTime":"2025-12-02T10:19:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.078336 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.078381 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.078396 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.078411 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.078422 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.181412 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.181442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.181451 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.181464 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.181474 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.283966 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.284012 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.284024 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.284038 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.284048 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.386663 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.386703 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.386711 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.386724 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.386732 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.489736 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.489791 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.489805 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.489821 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.489831 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.592499 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.592771 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.592908 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.593034 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.593192 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.695919 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.695964 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.695974 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.695988 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.695998 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.797955 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.797990 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.797999 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.798014 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.798023 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.900300 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.900355 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.900363 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.900378 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:53 crc kubenswrapper[4679]: I1202 10:19:53.900391 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:53Z","lastTransitionTime":"2025-12-02T10:19:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.002386 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.002451 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.002473 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.002502 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.002525 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.105350 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.105411 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.105433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.105462 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.105484 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.207671 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.207736 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.207753 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.207775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.207792 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.310358 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.310415 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.310434 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.310460 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.310477 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.413148 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.413194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.413202 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.413215 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.413224 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.515435 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.515477 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.515491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.515504 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.515515 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.618124 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.618160 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.618184 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.618196 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.618205 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.720631 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.720671 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.720681 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.720694 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.720705 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.823397 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.823690 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.823772 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.823850 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.823936 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.907685 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.907769 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.907721 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.907713 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:54 crc kubenswrapper[4679]: E1202 10:19:54.907840 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:54 crc kubenswrapper[4679]: E1202 10:19:54.907940 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:54 crc kubenswrapper[4679]: E1202 10:19:54.908108 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:54 crc kubenswrapper[4679]: E1202 10:19:54.908261 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.927495 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.927550 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.927579 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.927597 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.927611 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:54Z","lastTransitionTime":"2025-12-02T10:19:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:54 crc kubenswrapper[4679]: I1202 10:19:54.993148 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:54 crc kubenswrapper[4679]: E1202 10:19:54.993358 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:54 crc kubenswrapper[4679]: E1202 10:19:54.993424 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:20:10.993407006 +0000 UTC m=+64.323545876 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.029602 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.029679 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.029692 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.029707 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.029718 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.131685 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.131708 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.131716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.131729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.131737 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.234554 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.234591 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.234603 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.234620 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.234633 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.336593 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.336675 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.336699 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.336723 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.336739 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.440286 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.440397 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.440413 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.440431 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.440449 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.543908 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.544399 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.544538 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.544696 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.544821 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.648463 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.648817 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.648889 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.648969 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.649041 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.752660 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.753924 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.754040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.754196 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.754267 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.857467 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.857495 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.857506 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.857521 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.857530 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.959899 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.959938 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.959947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.959963 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:55 crc kubenswrapper[4679]: I1202 10:19:55.959974 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:55Z","lastTransitionTime":"2025-12-02T10:19:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.062879 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.062915 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.062925 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.062946 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.062956 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.165296 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.165379 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.165395 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.165416 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.165431 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.267615 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.267657 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.267671 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.267691 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.267705 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.370566 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.370598 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.370609 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.370626 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.370638 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.473527 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.473577 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.473587 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.473604 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.473614 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.576212 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.576248 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.576258 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.576274 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.576284 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.679093 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.679161 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.679174 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.679193 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.679225 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.782036 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.782999 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.783084 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.783187 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.783266 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.811776 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.811985 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.812021 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:20:28.811985598 +0000 UTC m=+82.142124508 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.812127 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.812140 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.812361 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:20:28.812282326 +0000 UTC m=+82.142421216 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.812404 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.812494 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:20:28.812471762 +0000 UTC m=+82.142610652 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.886228 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.886281 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.886296 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.886339 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.886352 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.908012 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.908062 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.908112 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.908026 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.908161 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.908243 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.908358 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:56 crc kubenswrapper[4679]: E1202 10:19:56.908415 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.920404 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.931037 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.942413 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.954247 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.970587 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.984460 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.988190 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.988218 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.988227 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.988240 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.988249 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:56Z","lastTransitionTime":"2025-12-02T10:19:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:56 crc kubenswrapper[4679]: I1202 10:19:56.995533 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:56Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.012738 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.013085 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.013127 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013233 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013250 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013260 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013278 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013295 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013332 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013317 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:20:29.013291966 +0000 UTC m=+82.343430826 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:57 crc kubenswrapper[4679]: E1202 10:19:57.013413 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:20:29.013380689 +0000 UTC m=+82.343519539 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.027200 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.048050 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.060927 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.074072 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.086589 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.092760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.092784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.092793 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.092805 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.092815 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.101851 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.113220 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.130735 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:57Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.194456 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.194505 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.194516 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.194530 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.194540 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.297055 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.297108 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.297119 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.297137 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.297148 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.399665 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.399705 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.399713 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.399727 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.399757 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.501828 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.501883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.501900 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.501922 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.501939 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.604895 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.604939 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.604948 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.604963 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.604974 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.707202 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.707248 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.707260 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.707277 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.707323 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.809273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.809349 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.809367 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.809389 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.809407 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.912013 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.912063 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.912075 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.912092 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:57 crc kubenswrapper[4679]: I1202 10:19:57.912104 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:57Z","lastTransitionTime":"2025-12-02T10:19:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.015204 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.015247 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.015258 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.015272 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.015281 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.117007 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.117053 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.117068 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.117086 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.117099 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.219493 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.219540 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.219551 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.219566 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.219576 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.322138 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.322194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.322208 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.322226 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.322239 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.347001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.347041 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.347051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.347065 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.347075 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.361951 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:58Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.365431 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.365486 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.365496 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.365509 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.365517 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.381758 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:58Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.385526 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.385568 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.385579 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.385594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.385605 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.400203 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:58Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.404177 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.404216 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.404231 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.404251 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.404265 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.421192 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:58Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.425596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.425638 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.425646 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.425661 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.425671 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.437800 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:19:58Z is after 2025-08-24T17:21:41Z" Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.437980 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.439906 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.439962 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.439973 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.439987 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.439997 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.542478 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.542517 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.542525 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.542539 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.542547 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.645094 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.645439 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.645531 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.645622 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.645704 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.748239 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.748288 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.748298 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.748333 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.748368 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.851166 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.852188 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.852391 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.852678 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.852823 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.908012 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.908326 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.908241 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.908102 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.908647 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.908817 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.908996 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:19:58 crc kubenswrapper[4679]: E1202 10:19:58.909130 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.955710 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.955751 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.955761 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.955774 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:58 crc kubenswrapper[4679]: I1202 10:19:58.955783 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:58Z","lastTransitionTime":"2025-12-02T10:19:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.058268 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.058342 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.058351 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.058367 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.058379 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.160486 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.160523 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.160532 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.160546 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.160555 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.263051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.263108 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.263117 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.263130 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.263139 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.365026 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.365082 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.365091 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.365106 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.365118 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.467518 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.467722 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.467738 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.467753 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.467765 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.570053 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.570085 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.570095 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.570109 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.570118 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.672951 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.673007 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.673023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.673042 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.673061 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.775265 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.775324 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.775333 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.775345 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.775354 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.878109 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.878160 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.878175 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.878194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.878211 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.908885 4679 scope.go:117] "RemoveContainer" containerID="b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.980735 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.981034 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.981044 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.981059 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:19:59 crc kubenswrapper[4679]: I1202 10:19:59.981069 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:19:59Z","lastTransitionTime":"2025-12-02T10:19:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.083967 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.084008 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.084020 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.084035 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.084046 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.186474 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.186536 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.186545 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.186560 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.186569 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.206340 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/1.log" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.208999 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.209378 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.232055 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.250233 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.270198 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.286549 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.290922 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.290960 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.290974 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.291191 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.291220 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.303844 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.315329 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.328568 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.341437 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.353778 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.366982 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.381991 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.394437 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.394830 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.394856 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.394876 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.394890 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.403850 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.414667 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.425859 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.437809 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.451983 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.496675 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.496717 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.496728 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.496742 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.496753 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.599241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.599289 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.599381 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.599409 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.599419 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.701649 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.701860 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.701946 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.702102 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.702207 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.804401 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.804466 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.804489 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.804515 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.804532 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.906407 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.907571 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.907750 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.907857 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.907888 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.908170 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:00 crc kubenswrapper[4679]: E1202 10:20:00.908284 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:00 crc kubenswrapper[4679]: E1202 10:20:00.908044 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.908159 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:00Z","lastTransitionTime":"2025-12-02T10:20:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.907847 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:00 crc kubenswrapper[4679]: I1202 10:20:00.907890 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:00 crc kubenswrapper[4679]: E1202 10:20:00.908523 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:00 crc kubenswrapper[4679]: E1202 10:20:00.908579 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.011647 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.011973 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.012185 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.012433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.012631 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.116000 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.116041 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.116052 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.116068 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.116079 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.214624 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/2.log" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.215816 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/1.log" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218297 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218577 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218643 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218706 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218760 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218906 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73" exitCode=1 Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.218963 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.219048 4679 scope.go:117] "RemoveContainer" containerID="b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.219554 4679 scope.go:117] "RemoveContainer" containerID="9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73" Dec 02 10:20:01 crc kubenswrapper[4679]: E1202 10:20:01.219782 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.236292 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.250663 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.268543 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.282603 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.297239 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.316025 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.321182 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.321219 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.321230 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.321246 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.321258 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.330568 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.342229 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.354663 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.368106 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.382953 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.396755 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.407532 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.417277 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.424029 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.424063 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.424075 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.424091 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.424103 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.429674 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.440159 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:01Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.526459 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.526499 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.526507 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.526535 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.526544 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.629328 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.629398 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.629411 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.629430 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.629440 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.731292 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.731353 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.731363 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.731379 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.731390 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.833835 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.833877 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.833886 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.833902 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.833915 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.936277 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.936337 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.936347 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.936362 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:01 crc kubenswrapper[4679]: I1202 10:20:01.936371 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:01Z","lastTransitionTime":"2025-12-02T10:20:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.038296 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.038353 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.038364 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.038378 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.038386 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.141461 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.141505 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.141517 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.141534 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.141545 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.224732 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/2.log" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.225431 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.228519 4679 scope.go:117] "RemoveContainer" containerID="9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73" Dec 02 10:20:02 crc kubenswrapper[4679]: E1202 10:20:02.228701 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.238741 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.242408 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.244181 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.244286 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.244365 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.244443 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.244513 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.256635 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.271184 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.286661 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.304540 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.316766 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.331098 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.342426 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.346925 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.346973 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.346984 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.347000 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.347011 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.352617 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.363291 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.375154 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.387065 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.400939 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.418425 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.432535 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.448930 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.448971 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.448981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.448996 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.449009 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.459646 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3e672cd93aee5ff255742c18fb07c87c5f25c1ab12914803269a8c7443459da\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"message\\\":\\\"ing-signer@1740288168] [{operator.openshift.io/v1 DNS default d8d88c7e-8c3e-49b6-8c5b-84aa454da2d7 0xc007748ad7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:dns,Protocol:UDP,Port:53,TargetPort:{1 0 dns},NodePort:0,AppProtocol:nil,},ServicePort{Name:dns-tcp,Protocol:TCP,Port:53,TargetPort:{1 0 dns-tcp},NodePort:0,AppProtocol:nil,},ServicePort{Name:metrics,Protocol:TCP,Port:9154,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{dns.operator.openshift.io/daemonset-dns: default,},ClusterIP:10.217.4.10,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.10],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1202 10:19:36.272683 6078 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initia\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.472903 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.488995 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.501350 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.513949 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.526528 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.538449 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.554269 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.554322 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.554334 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.554350 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.554361 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.554952 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.567033 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.579213 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.590361 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.601645 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.617281 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.638570 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.651897 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.656610 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.656649 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.656658 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.656673 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.656684 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.665869 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.676911 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.688126 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:02Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.759144 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.759656 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.759733 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.759805 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.759868 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.862547 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.862792 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.862874 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.862956 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.863029 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.908649 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.908666 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.908813 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:02 crc kubenswrapper[4679]: E1202 10:20:02.909293 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:02 crc kubenswrapper[4679]: E1202 10:20:02.909123 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:02 crc kubenswrapper[4679]: E1202 10:20:02.909390 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.908863 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:02 crc kubenswrapper[4679]: E1202 10:20:02.909459 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.965716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.965755 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.965764 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.965784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:02 crc kubenswrapper[4679]: I1202 10:20:02.965795 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:02Z","lastTransitionTime":"2025-12-02T10:20:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.067846 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.068117 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.068221 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.068374 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.068497 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.171557 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.171599 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.171608 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.171636 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.171648 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.274472 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.274531 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.274547 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.274571 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.274589 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.376560 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.376603 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.376614 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.376629 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.376642 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.478817 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.478863 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.478876 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.478890 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.478903 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.581554 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.581596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.581608 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.581636 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.581646 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.683801 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.683845 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.683856 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.683870 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.683881 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.786267 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.786336 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.786349 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.786367 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.786380 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.888971 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.889029 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.889040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.889056 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.889066 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.990753 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.990788 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.990798 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.990811 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:03 crc kubenswrapper[4679]: I1202 10:20:03.990821 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:03Z","lastTransitionTime":"2025-12-02T10:20:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.093259 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.093319 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.093329 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.093344 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.093356 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.195781 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.195827 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.195840 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.195858 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.195870 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.297972 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.298020 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.298032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.298051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.298062 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.400565 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.400606 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.400616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.400631 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.400642 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.503717 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.503776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.503787 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.503805 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.503820 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.608114 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.608622 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.608650 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.608672 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.608685 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.710918 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.710948 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.710956 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.710968 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.710976 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.813864 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.814102 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.814168 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.814267 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.814354 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.908249 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.908294 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.908411 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:04 crc kubenswrapper[4679]: E1202 10:20:04.908500 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.908517 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:04 crc kubenswrapper[4679]: E1202 10:20:04.908668 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:04 crc kubenswrapper[4679]: E1202 10:20:04.908694 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:04 crc kubenswrapper[4679]: E1202 10:20:04.908855 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.915955 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.916006 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.916023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.916047 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:04 crc kubenswrapper[4679]: I1202 10:20:04.916066 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:04Z","lastTransitionTime":"2025-12-02T10:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.019164 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.019237 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.019262 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.019292 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.019357 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.122121 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.122196 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.122217 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.122241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.122259 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.224664 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.224987 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.225009 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.225024 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.225035 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.327933 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.327976 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.327988 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.328005 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.328019 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.430015 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.430051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.430061 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.430078 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.430092 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.532679 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.532732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.532751 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.532773 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.532790 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.635228 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.635283 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.635294 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.635383 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.635407 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.738246 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.738293 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.738318 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.738338 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.738362 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.841189 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.841228 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.841239 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.841254 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.841266 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.943400 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.943453 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.943464 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.943482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:05 crc kubenswrapper[4679]: I1202 10:20:05.943492 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:05Z","lastTransitionTime":"2025-12-02T10:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.045815 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.045861 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.045872 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.045887 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.045897 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.148885 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.148977 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.149000 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.149023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.149037 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.251069 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.251103 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.251115 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.251130 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.251139 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.353601 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.353661 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.353677 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.353698 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.353714 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.456273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.456440 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.456481 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.456511 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.456534 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.558953 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.558990 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.558999 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.559012 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.559024 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.661079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.661116 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.661128 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.661143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.661154 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.763203 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.763241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.763251 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.763263 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.763272 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.868415 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.869596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.869614 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.869632 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.869646 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.907656 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.907720 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.907740 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:06 crc kubenswrapper[4679]: E1202 10:20:06.907914 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.907963 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:06 crc kubenswrapper[4679]: E1202 10:20:06.908014 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:06 crc kubenswrapper[4679]: E1202 10:20:06.908064 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:06 crc kubenswrapper[4679]: E1202 10:20:06.908707 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.922648 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:06Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.943812 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:06Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.955538 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:06Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.970285 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:06Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.971838 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.971894 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.971933 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.971954 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.971971 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:06Z","lastTransitionTime":"2025-12-02T10:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:06 crc kubenswrapper[4679]: I1202 10:20:06.982408 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:06Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.012530 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.045726 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.058866 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.070092 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.074039 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.074116 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.074127 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.074143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.074155 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.082385 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.091019 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.100333 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.110813 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.123359 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.135945 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.147650 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.155985 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:07Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.175917 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.175959 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.175971 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.175997 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.176010 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.278358 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.278409 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.278419 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.278436 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.278449 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.381388 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.381428 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.381439 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.381454 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.381465 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.484012 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.484652 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.484689 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.484710 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.484720 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.587383 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.587452 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.587466 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.587483 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.587494 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.689642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.689694 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.689715 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.689742 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.689764 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.792161 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.792238 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.792248 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.792261 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.792270 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.894474 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.894516 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.894528 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.894544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.894557 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.996743 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.996787 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.996804 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.996827 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:07 crc kubenswrapper[4679]: I1202 10:20:07.996845 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:07Z","lastTransitionTime":"2025-12-02T10:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.099581 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.099626 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.099641 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.099663 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.099699 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.201993 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.202028 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.202058 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.202074 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.202084 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.304626 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.304662 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.304670 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.304686 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.304695 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.407467 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.407510 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.407522 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.407537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.407548 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.510729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.510773 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.510783 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.510802 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.510813 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.565880 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.565926 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.565936 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.565953 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.565966 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.579930 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:08Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.583171 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.583201 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.583210 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.583241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.583249 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.595767 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:08Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.599391 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.599421 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.599429 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.599441 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.599451 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.612132 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:08Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.615273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.615318 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.615331 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.615348 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.615360 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.626675 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:08Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.630669 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.630709 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.630724 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.630744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.630759 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.646600 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:08Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.647025 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.648920 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.648983 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.649001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.649592 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.649675 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.753687 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.753776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.753788 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.753804 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.753814 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.856780 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.856818 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.856829 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.856842 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.856854 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.907739 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.907800 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.907843 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.908095 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.907940 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.908182 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.908239 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:08 crc kubenswrapper[4679]: E1202 10:20:08.908385 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.959430 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.959463 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.959471 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.959485 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:08 crc kubenswrapper[4679]: I1202 10:20:08.959495 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:08Z","lastTransitionTime":"2025-12-02T10:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.062553 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.062616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.062633 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.062655 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.062671 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.165237 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.165275 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.165285 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.165325 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.165339 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.267883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.267917 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.267983 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.268004 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.268017 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.370991 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.371034 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.371050 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.371070 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.371091 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.473141 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.473172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.473180 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.473192 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.473201 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.575706 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.575745 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.575755 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.575769 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.575780 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.678120 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.678170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.678185 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.678203 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.678215 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.780853 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.780914 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.780929 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.780954 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.780970 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.884340 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.884408 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.884422 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.884439 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.884452 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.924581 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.986687 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.986738 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.986769 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.986789 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:09 crc kubenswrapper[4679]: I1202 10:20:09.986801 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:09Z","lastTransitionTime":"2025-12-02T10:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.089673 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.089712 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.089720 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.089737 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.089748 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.192399 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.192471 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.192491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.192517 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.192539 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.295244 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.295282 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.295291 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.295318 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.295328 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.397749 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.397792 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.397803 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.397818 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.397831 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.501126 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.501178 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.501188 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.501206 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.501220 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.603735 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.603776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.603784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.603797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.603805 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.706290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.706345 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.706352 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.706365 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.706373 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.807983 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.808023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.808032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.808046 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.808056 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.908672 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.908881 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:10 crc kubenswrapper[4679]: E1202 10:20:10.909069 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.909288 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.909337 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:10 crc kubenswrapper[4679]: E1202 10:20:10.909405 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:10 crc kubenswrapper[4679]: E1202 10:20:10.909485 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:10 crc kubenswrapper[4679]: E1202 10:20:10.909671 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.910082 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.910110 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.910121 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.910137 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:10 crc kubenswrapper[4679]: I1202 10:20:10.910152 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:10Z","lastTransitionTime":"2025-12-02T10:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.012792 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.012825 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.012834 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.012851 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.012862 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.056116 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:11 crc kubenswrapper[4679]: E1202 10:20:11.056277 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:20:11 crc kubenswrapper[4679]: E1202 10:20:11.056353 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:20:43.056333494 +0000 UTC m=+96.386472354 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.115040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.115078 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.115089 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.115104 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.115115 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.218054 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.218090 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.218100 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.218116 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.218127 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.320873 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.320913 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.320923 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.320939 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.320951 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.423930 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.423988 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.424001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.424019 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.424033 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.526822 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.526900 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.526919 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.526947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.526964 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.629723 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.629768 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.629779 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.629792 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.629802 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.732736 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.732788 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.732800 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.732813 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.732822 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.835940 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.835974 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.835982 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.835997 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.836006 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.938473 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.938515 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.938526 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.938542 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:11 crc kubenswrapper[4679]: I1202 10:20:11.938554 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:11Z","lastTransitionTime":"2025-12-02T10:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.041057 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.041092 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.041102 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.041117 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.041127 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.144097 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.144163 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.144178 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.144199 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.144207 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.247907 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.247983 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.248001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.248033 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.248055 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.260297 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/0.log" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.260388 4679 generic.go:334] "Generic (PLEG): container finished" podID="06f77afa-f710-437b-9cfe-10959bdc1ac8" containerID="e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876" exitCode=1 Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.260434 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerDied","Data":"e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.260982 4679 scope.go:117] "RemoveContainer" containerID="e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.273157 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.288693 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.302382 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.316466 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.333907 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.349264 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.350988 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.351046 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.351060 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.351079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.351113 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.362504 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.376554 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.395852 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.406509 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.417589 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.432207 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.446201 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.452985 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.453015 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.453023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.453037 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.453048 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.460178 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.472933 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.483814 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.496805 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.510030 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:12Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.556246 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.556347 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.556364 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.556386 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.556402 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.658806 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.658849 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.658859 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.658875 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.658887 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.761653 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.761702 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.761714 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.761733 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.761743 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.864395 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.864473 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.864485 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.864502 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.864513 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.908395 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.908434 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.908469 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:12 crc kubenswrapper[4679]: E1202 10:20:12.908557 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:12 crc kubenswrapper[4679]: E1202 10:20:12.908650 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:12 crc kubenswrapper[4679]: E1202 10:20:12.908688 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.908737 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:12 crc kubenswrapper[4679]: E1202 10:20:12.908806 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.967407 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.967452 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.967461 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.967476 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:12 crc kubenswrapper[4679]: I1202 10:20:12.967486 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:12Z","lastTransitionTime":"2025-12-02T10:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.069881 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.069923 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.069932 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.069950 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.069963 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.172127 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.172178 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.172193 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.172213 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.172227 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.265979 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/0.log" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.266044 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerStarted","Data":"97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.275503 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.275750 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.275836 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.275921 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.275995 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.281136 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.297821 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.312588 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.325580 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.337387 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.349361 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.360479 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.371941 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.378750 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.378797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.378811 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.378830 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.378844 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.384423 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.394743 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.407101 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.420088 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.430506 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.441650 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.452412 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.468417 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.479462 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.481128 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.481241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.481275 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.481296 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.481324 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.494041 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:13Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.584192 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.584239 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.584253 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.584273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.584290 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.687367 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.687410 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.687422 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.687437 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.687447 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.790670 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.790724 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.790737 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.790756 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.790768 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.893912 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.893972 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.893982 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.893997 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.894028 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:13Z","lastTransitionTime":"2025-12-02T10:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:13 crc kubenswrapper[4679]: I1202 10:20:13.909051 4679 scope.go:117] "RemoveContainer" containerID="9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73" Dec 02 10:20:13 crc kubenswrapper[4679]: E1202 10:20:13.909277 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.005190 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.005241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.005262 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.005283 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.005298 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.108968 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.109021 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.109030 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.109045 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.109058 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.212129 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.212175 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.212186 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.212201 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.212212 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.314606 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.314643 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.314656 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.314673 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.314684 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.417260 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.417334 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.417346 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.417362 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.417372 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.519407 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.519447 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.519457 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.519472 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.519483 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.622176 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.622216 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.622229 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.622246 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.622258 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.724745 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.724782 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.724792 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.724806 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.724815 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.827631 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.827673 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.827692 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.827713 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.827730 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.907591 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.907664 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:14 crc kubenswrapper[4679]: E1202 10:20:14.907696 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:14 crc kubenswrapper[4679]: E1202 10:20:14.907798 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.907853 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:14 crc kubenswrapper[4679]: E1202 10:20:14.907898 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.908003 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:14 crc kubenswrapper[4679]: E1202 10:20:14.908053 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.929457 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.929485 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.929493 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.929503 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:14 crc kubenswrapper[4679]: I1202 10:20:14.929512 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:14Z","lastTransitionTime":"2025-12-02T10:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.031324 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.031361 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.031377 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.031391 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.031401 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.133354 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.133406 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.133416 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.133431 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.133440 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.235404 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.235438 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.235447 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.235459 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.235468 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.337613 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.337700 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.337716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.337737 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.337752 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.440536 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.440596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.440618 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.440645 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.440666 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.543429 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.543751 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.543889 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.544021 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.544168 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.646366 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.646715 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.646885 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.647029 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.647249 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.750596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.750634 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.750642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.750655 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.750664 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.853542 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.853832 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.853911 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.853995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.854068 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.956457 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.956513 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.956529 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.956550 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:15 crc kubenswrapper[4679]: I1202 10:20:15.956568 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:15Z","lastTransitionTime":"2025-12-02T10:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.058990 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.059039 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.059054 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.059075 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.059089 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.161012 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.161342 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.161431 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.161539 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.161634 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.264156 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.264465 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.264553 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.264661 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.264748 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.367005 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.367563 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.367659 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.367726 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.367794 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.469811 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.470040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.470126 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.470222 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.470291 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.573109 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.573153 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.573168 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.573189 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.573203 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.674853 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.675111 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.675194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.675271 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.675365 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.778096 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.778139 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.778156 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.778177 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.778190 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.886017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.886068 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.886081 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.886100 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.886111 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.908544 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.908719 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:16 crc kubenswrapper[4679]: E1202 10:20:16.908883 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:16 crc kubenswrapper[4679]: E1202 10:20:16.908757 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.909092 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:16 crc kubenswrapper[4679]: E1202 10:20:16.909185 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.909435 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:16 crc kubenswrapper[4679]: E1202 10:20:16.909706 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.955874 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:16Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.975409 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:16Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.988404 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.988443 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.988456 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.988475 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.988487 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:16Z","lastTransitionTime":"2025-12-02T10:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:16 crc kubenswrapper[4679]: I1202 10:20:16.989490 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:16Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.004948 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.028942 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.046944 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.059445 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.070139 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.082686 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.094875 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.095848 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.095980 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.096059 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.096228 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.096338 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.107469 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.115963 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.126151 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.139787 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.149442 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.159725 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.169604 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.180181 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:17Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.199240 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.199501 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.199619 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.199711 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.199794 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.302642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.302695 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.302707 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.302724 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.302757 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.405046 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.405091 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.405102 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.405118 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.405127 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.507373 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.507420 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.507431 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.507448 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.507458 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.609908 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.609953 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.609968 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.609985 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.609995 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.712094 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.712138 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.712150 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.712168 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.712181 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.814986 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.815035 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.815046 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.815060 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.815070 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.917136 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.917167 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.917176 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.917190 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:17 crc kubenswrapper[4679]: I1202 10:20:17.917198 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:17Z","lastTransitionTime":"2025-12-02T10:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.019343 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.019382 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.019394 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.019411 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.019421 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.122736 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.122777 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.122786 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.122799 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.122808 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.225128 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.225164 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.225172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.225185 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.225194 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.327809 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.327855 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.327868 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.327885 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.327897 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.430086 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.430121 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.430129 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.430140 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.430149 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.532448 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.532482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.532490 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.532504 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.532514 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.634875 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.634915 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.634926 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.634939 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.634950 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.737058 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.737092 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.737106 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.737122 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.737132 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.839338 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.839610 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.839695 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.839783 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.839867 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.908699 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.908794 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:18 crc kubenswrapper[4679]: E1202 10:20:18.908833 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.908704 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.908885 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:18 crc kubenswrapper[4679]: E1202 10:20:18.908977 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:18 crc kubenswrapper[4679]: E1202 10:20:18.909022 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:18 crc kubenswrapper[4679]: E1202 10:20:18.909071 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.942052 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.942356 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.942449 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.942526 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.942582 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.971359 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.971392 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.971401 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.971414 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.971424 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:18 crc kubenswrapper[4679]: E1202 10:20:18.989996 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:18Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.993504 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.993535 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.993544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.993557 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:18 crc kubenswrapper[4679]: I1202 10:20:18.993565 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:18Z","lastTransitionTime":"2025-12-02T10:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: E1202 10:20:19.004534 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:19Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.007595 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.007636 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.007646 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.007661 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.007672 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: E1202 10:20:19.017911 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:19Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.021172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.021208 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.021215 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.021229 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.021238 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: E1202 10:20:19.031589 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:19Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.034895 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.035064 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.035331 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.035516 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.035591 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: E1202 10:20:19.047754 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:19Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:19 crc kubenswrapper[4679]: E1202 10:20:19.047873 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.049623 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.049649 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.049660 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.049674 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.049683 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.152148 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.152195 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.152205 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.152222 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.152232 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.254689 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.254862 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.254885 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.254914 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.254945 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.357253 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.357293 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.357318 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.357331 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.357342 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.459128 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.459161 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.459171 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.459184 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.459193 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.561834 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.561895 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.561908 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.561926 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.561938 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.664112 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.664144 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.664152 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.664177 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.664186 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.767081 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.767113 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.767125 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.767143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.767155 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.870091 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.870147 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.870157 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.870171 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.870180 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.972548 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.972587 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.972596 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.972610 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:19 crc kubenswrapper[4679]: I1202 10:20:19.972619 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:19Z","lastTransitionTime":"2025-12-02T10:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.075277 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.075359 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.075371 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.075388 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.075400 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.177909 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.177947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.177956 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.177970 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.177979 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.279959 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.280032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.280043 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.280056 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.280066 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.382683 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.382745 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.382758 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.382776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.382787 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.484947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.485176 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.485187 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.485201 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.485213 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.587643 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.587697 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.587713 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.587732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.587752 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.690601 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.690646 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.690657 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.690672 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.690682 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.793237 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.793284 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.793329 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.793351 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.793370 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.896017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.896056 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.896064 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.896078 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.896088 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.908228 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.908290 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.908247 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:20 crc kubenswrapper[4679]: E1202 10:20:20.908372 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:20 crc kubenswrapper[4679]: E1202 10:20:20.908532 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.908560 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:20 crc kubenswrapper[4679]: E1202 10:20:20.908674 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:20 crc kubenswrapper[4679]: E1202 10:20:20.908761 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.998113 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.998148 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.998156 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.998170 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:20 crc kubenswrapper[4679]: I1202 10:20:20.998180 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:20Z","lastTransitionTime":"2025-12-02T10:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.099961 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.099995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.100003 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.100015 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.100025 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.202337 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.202382 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.202391 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.202410 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.202420 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.304108 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.304161 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.304173 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.304190 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.304202 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.406821 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.406865 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.406877 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.406891 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.406900 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.509395 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.509437 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.509447 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.509464 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.509474 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.612666 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.612704 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.612713 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.612728 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.612738 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.715608 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.715655 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.715663 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.715676 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.715685 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.817682 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.817722 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.817732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.817747 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.817755 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.919750 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.919795 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.919805 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.919827 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:21 crc kubenswrapper[4679]: I1202 10:20:21.919840 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:21Z","lastTransitionTime":"2025-12-02T10:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.022482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.022529 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.022540 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.022554 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.022565 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.124879 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.124922 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.124932 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.124947 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.124957 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.227296 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.227393 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.227416 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.227463 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.227487 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.330583 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.330647 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.330664 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.330688 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.330709 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.433440 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.433513 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.433537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.433565 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.433588 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.536712 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.536763 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.536781 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.536803 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.536821 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.640154 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.640189 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.640197 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.640210 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.640219 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.742150 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.742203 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.742211 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.742225 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.742233 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.844482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.844547 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.844558 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.844574 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.844586 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.907723 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.907761 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.907759 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:22 crc kubenswrapper[4679]: E1202 10:20:22.907950 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.908001 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:22 crc kubenswrapper[4679]: E1202 10:20:22.908077 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:22 crc kubenswrapper[4679]: E1202 10:20:22.908234 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:22 crc kubenswrapper[4679]: E1202 10:20:22.908786 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.948103 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.948143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.948156 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.948206 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:22 crc kubenswrapper[4679]: I1202 10:20:22.948218 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:22Z","lastTransitionTime":"2025-12-02T10:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.050589 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.050695 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.050729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.050759 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.050781 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.152514 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.152560 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.152570 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.152585 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.152595 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.254663 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.254714 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.254724 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.254743 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.254753 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.356938 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.356989 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.357001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.357017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.357028 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.459016 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.459070 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.459094 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.459114 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.459129 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.561570 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.561605 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.561616 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.561633 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.561644 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.663895 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.663954 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.663970 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.663995 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.664013 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.767090 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.767127 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.767136 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.767151 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.767162 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.870445 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.870483 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.870493 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.870508 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.870533 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.973688 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.973723 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.973734 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.973749 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:23 crc kubenswrapper[4679]: I1202 10:20:23.973760 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:23Z","lastTransitionTime":"2025-12-02T10:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.075680 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.075738 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.075755 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.075779 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.075795 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.178839 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.178900 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.178919 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.178946 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.178963 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.282477 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.282550 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.282567 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.282593 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.282611 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.385274 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.385369 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.385382 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.385410 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.385429 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.488504 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.488557 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.488571 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.488585 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.488595 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.592166 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.592233 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.592247 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.592273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.592288 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.695183 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.695236 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.695244 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.695262 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.695271 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.797779 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.797830 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.797845 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.797899 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.797915 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.900529 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.900567 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.900578 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.900594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.900606 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:24Z","lastTransitionTime":"2025-12-02T10:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.908087 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.908154 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:24 crc kubenswrapper[4679]: E1202 10:20:24.908188 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.908256 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:24 crc kubenswrapper[4679]: I1202 10:20:24.908090 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:24 crc kubenswrapper[4679]: E1202 10:20:24.908664 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:24 crc kubenswrapper[4679]: E1202 10:20:24.908822 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:24 crc kubenswrapper[4679]: E1202 10:20:24.908886 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.002335 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.002380 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.002391 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.002454 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.002469 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.107870 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.107922 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.107935 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.107956 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.107970 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.210535 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.210579 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.210593 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.210608 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.210617 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.313019 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.313073 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.313085 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.313103 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.313115 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.414908 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.414949 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.414960 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.414975 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.414989 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.517090 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.517123 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.517131 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.517149 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.517157 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.620642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.621297 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.621414 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.621499 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.621632 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.724933 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.725051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.725072 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.725095 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.725111 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.827279 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.827366 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.827383 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.827404 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.827421 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.908767 4679 scope.go:117] "RemoveContainer" containerID="9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.929772 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.929797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.929806 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.929818 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:25 crc kubenswrapper[4679]: I1202 10:20:25.929826 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:25Z","lastTransitionTime":"2025-12-02T10:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.031950 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.032001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.032018 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.032044 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.032062 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.134226 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.134264 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.134273 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.134287 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.134298 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.236493 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.236535 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.236546 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.236564 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.236576 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.338971 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.339073 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.339089 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.339145 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.339162 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.442454 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.442498 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.442510 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.442529 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.442541 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.545032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.545066 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.545075 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.545087 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.545096 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.647961 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.648016 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.648085 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.648108 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.648123 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.750026 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.750059 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.750069 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.750094 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.750104 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.852732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.852804 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.852820 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.852839 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.852854 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.908564 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:26 crc kubenswrapper[4679]: E1202 10:20:26.908985 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.908625 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:26 crc kubenswrapper[4679]: E1202 10:20:26.909058 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.908696 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.908623 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:26 crc kubenswrapper[4679]: E1202 10:20:26.909122 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:26 crc kubenswrapper[4679]: E1202 10:20:26.909471 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.923458 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.944551 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.956820 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.956913 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.956932 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.956951 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.956966 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:26Z","lastTransitionTime":"2025-12-02T10:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.966396 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:26 crc kubenswrapper[4679]: I1202 10:20:26.986098 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:26Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.003594 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.019572 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.035351 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.049372 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.059537 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.059588 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.059602 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.059622 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.059635 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.063670 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.082250 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.128187 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.143945 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.157873 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.162806 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.162854 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.162866 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.162883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.162899 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.176327 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.192769 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.203963 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.216421 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.228537 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.264197 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.264230 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.264241 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.264256 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.264267 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.309777 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/2.log" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.312028 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.313069 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.326377 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.340586 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.354139 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.365521 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.366984 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.367132 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.367196 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.367254 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.367336 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.377329 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.392637 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.406608 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.421502 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.436139 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.455005 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.465641 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.469433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.469471 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.469480 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.469493 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.469504 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.477583 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.489290 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.502188 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.514693 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.531357 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.545400 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.556493 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:27Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.572649 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.572696 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.572709 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.572728 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.572740 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.675037 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.675090 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.675104 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.675121 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.675132 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.777776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.777813 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.777822 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.777836 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.777845 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.880676 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.880754 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.880771 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.880797 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.880814 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.983798 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.984084 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.984159 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.984231 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:27 crc kubenswrapper[4679]: I1202 10:20:27.984293 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:27Z","lastTransitionTime":"2025-12-02T10:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.086955 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.087004 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.087017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.087034 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.087045 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.189099 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.189140 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.189151 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.189167 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.189179 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.291624 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.291663 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.291674 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.291688 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.291698 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.318890 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/3.log" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.319648 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/2.log" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.322028 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" exitCode=1 Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.322076 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.322122 4679 scope.go:117] "RemoveContainer" containerID="9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.322880 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.323139 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.341072 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.354347 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.370074 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.385278 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.394017 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.394038 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.394048 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.394063 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.394073 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.397208 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.412037 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.425643 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.436818 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.450917 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.463459 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.474825 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.486130 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.495682 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.496505 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.496523 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.496532 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.496544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.497263 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.506475 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.516873 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.527979 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.539400 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.553868 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e4bc4c7c752c944ce60c5a8bae0cfa8f600496be4f15ecf923d8036d330be73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:00Z\\\",\\\"message\\\":\\\"t(s)\\\\nF1202 10:20:00.643959 6356 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:00Z is after 2025-08-24T17:21:41Z]\\\\nI1202 10:20:00.643963 6356 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-lzf8q\\\\nI1202 10:20:00.643958 6356 services_controller.go:454] Service openshift-apiserver-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1202 10:20:00.643975 6356 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-7p\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:28Z\\\",\\\"message\\\":\\\"achine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1202 10:20:27.789985 6700 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 10:20:27.790019 6700 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:28Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.599831 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.599913 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.599924 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.599937 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.599946 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.702451 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.702482 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.702491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.702504 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.702513 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.804376 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.804437 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.804445 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.804460 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.804469 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.837007 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.837116 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.837139 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:32.837110542 +0000 UTC m=+146.167249422 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.837200 4679 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.837254 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.837261 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:21:32.837243146 +0000 UTC m=+146.167382076 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.837412 4679 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.837490 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 10:21:32.837479472 +0000 UTC m=+146.167618392 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.906822 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.906860 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.906870 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.906883 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.906892 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:28Z","lastTransitionTime":"2025-12-02T10:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.908035 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.908075 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.908035 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.908132 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:28 crc kubenswrapper[4679]: I1202 10:20:28.908152 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.908214 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.908426 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:28 crc kubenswrapper[4679]: E1202 10:20:28.908546 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.009522 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.009556 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.009566 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.009579 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.009588 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.040290 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.040370 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040502 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040521 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040531 4679 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040535 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040577 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 10:21:33.040564444 +0000 UTC m=+146.370703304 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040583 4679 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040612 4679 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.040718 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 10:21:33.040692728 +0000 UTC m=+146.370831628 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.112395 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.112442 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.112454 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.112472 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.112487 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.215290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.215389 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.215406 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.215430 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.215447 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.318845 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.318936 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.318955 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.318978 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.318996 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.328292 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/3.log" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.333046 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.333288 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.353679 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.376971 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://016acf92711489a1395ec1cba8986fb1522d1399839a9dd4ea4f6cbb832ce331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.392810 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a1d8d7-802a-4246-84b9-7cb1880de962\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"file observer\\\\nW1202 10:19:24.222100 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 10:19:24.222437 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 10:19:24.230759 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-185412186/tls.crt::/tmp/serving-cert-185412186/tls.key\\\\\\\"\\\\nI1202 10:19:24.639899 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 10:19:24.655755 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 10:19:24.655789 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 10:19:24.655823 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 10:19:24.655829 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 10:19:24.660517 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 10:19:24.660542 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660548 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 10:19:24.660553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 10:19:24.660557 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 10:19:24.660560 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 10:19:24.660564 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 10:19:24.660700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 10:19:24.661631 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.408708 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a0b357e-83e1-4ed7-ae2a-8803a3e6c8dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa439ef6fc541ae20fd276330cd2cbae0a408cb3b84b066bb159d7b0851060ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://703480e504f3109f86f7734bfa244c39db59cad51158a6ef096bca99ee9a019a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07484cab94d8fd7a3b6687afe184ef6eb0fc5be8482a5f28ffa7991947189532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fea14808345db81d0d7734e5fb23668c234f6ad13295e813cc167fe6e0391cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.418048 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.418148 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.418181 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.418214 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.418235 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.420532 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9464d7f0-62da-43d4-891c-6461a8cd82f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c469b3ffce082c2f1877464ffc8cb8d864322485d25b8fda557793c51425ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ab2d31f518041481902632b90894404d8215e15af3e8a03cb4fa30be3c05cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.430013 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.432048 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.436132 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.436167 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.436178 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.436194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.436207 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.442262 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.448105 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.451485 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.451513 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.451524 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.451540 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.451551 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.454339 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.464887 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.466577 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.470872 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.470916 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.470931 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.470952 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.470968 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.477851 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.483825 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.486385 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.486410 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.486418 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.486432 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.486443 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.490203 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.496915 4679 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"128cffb4-d188-4793-ba88-90c1cc2d9356\\\",\\\"systemUUID\\\":\\\"6eb2d3aa-2e30-4ebf-a13e-8a12a9f21c5f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: E1202 10:20:29.497557 4679 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.499005 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.499030 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.499039 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.499062 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.499072 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.500339 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.510974 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-r5sft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f77afa-f710-437b-9cfe-10959bdc1ac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:20:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:12Z\\\",\\\"message\\\":\\\"2025-12-02T10:19:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5\\\\n2025-12-02T10:19:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_2fb7f327-ab19-448e-bfc6-a62378d8f1c5 to /host/opt/cni/bin/\\\\n2025-12-02T10:19:26Z [verbose] multus-daemon started\\\\n2025-12-02T10:19:26Z [verbose] Readiness Indicator file check\\\\n2025-12-02T10:20:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:20:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6dq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-r5sft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.523164 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09d3af27-a3cd-4f88-a70f-941b22089a16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ff97142329fbf370b832f982a387ef14fb5ff57b7ab56cf03da2ad4e7cfd39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1d629a1897ee35e02aa723570551cdd8861dc92d2eaed8739b16e522a3c7347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2338cdd299ae00cc38da4c00dd8d1750618470ebc3af6d4ff32858d028580f4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a22d4c4f374dd5faae1069085da41b310611a5398c863f64d6d0744eaab913b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe602a197b5b0931ca9eb30905fcf5308f22fc628851e8c90c2c17ccda78b0ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bbaf724101dbf01e8285236041d886689823affcbece72471cec4e14c572d7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba13a358423c51f84ad4a60764fc29ff9d20c189d346b4ca2e60691cd8177b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpck8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2lmfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.542969 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a66c9e2c-2ca3-4348-84cc-19f365505c9e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T10:20:28Z\\\",\\\"message\\\":\\\"achine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1202 10:20:27.789985 6700 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 10:20:27.790019 6700 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T10:20:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T10:19:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cmbpj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pjb6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.552683 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.562507 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.572720 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:29Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.601776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.601870 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.601897 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.601940 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.601958 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.704718 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.704782 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.704804 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.704831 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.704851 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.807668 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.807745 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.807768 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.807800 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.807823 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.910172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.910233 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.910244 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.910256 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:29 crc kubenswrapper[4679]: I1202 10:20:29.910287 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:29Z","lastTransitionTime":"2025-12-02T10:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.012963 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.013005 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.013013 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.013025 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.013034 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.115949 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.116023 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.116047 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.116079 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.116102 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.219337 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.219377 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.219390 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.219406 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.219420 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.322826 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.322899 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.322926 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.322953 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.322974 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.425716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.425784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.425807 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.425835 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.425856 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.528578 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.528629 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.528646 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.528669 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.528685 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.632008 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.632063 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.632078 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.632108 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.632127 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.735013 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.735069 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.735081 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.735100 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.735141 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.837357 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.837407 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.837421 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.837440 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.837454 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.907911 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.907958 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.907919 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.908138 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:30 crc kubenswrapper[4679]: E1202 10:20:30.908220 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:30 crc kubenswrapper[4679]: E1202 10:20:30.908267 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:30 crc kubenswrapper[4679]: E1202 10:20:30.908394 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:30 crc kubenswrapper[4679]: E1202 10:20:30.908519 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.940511 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.940554 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.940564 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.940578 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:30 crc kubenswrapper[4679]: I1202 10:20:30.940589 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:30Z","lastTransitionTime":"2025-12-02T10:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.043290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.043342 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.043351 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.043366 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.043397 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.145112 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.145174 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.145189 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.145211 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.145228 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.247621 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.247672 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.247686 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.247702 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.247713 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.349708 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.349761 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.349771 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.349788 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.349800 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.452999 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.453074 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.453087 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.453113 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.453132 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.556093 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.556160 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.556172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.556193 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.556204 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.660497 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.660566 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.660582 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.660607 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.660630 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.762683 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.762732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.762742 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.762755 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.762765 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.865119 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.865158 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.865166 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.865179 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.865188 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.968360 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.968405 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.968423 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.968438 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:31 crc kubenswrapper[4679]: I1202 10:20:31.968448 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:31Z","lastTransitionTime":"2025-12-02T10:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.070725 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.070769 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.070784 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.070799 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.070810 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.174793 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.174846 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.174860 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.174877 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.174890 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.277172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.277216 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.277227 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.277245 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.277257 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.379692 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.379732 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.379743 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.379758 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.379769 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.482451 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.482491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.482502 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.482517 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.482530 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.584910 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.584946 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.584954 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.584968 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.584976 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.687340 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.687387 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.687398 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.687415 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.687427 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.789544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.789579 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.789589 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.789601 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.789609 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.891627 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.891673 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.891690 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.891705 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.891716 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.908015 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.908087 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:32 crc kubenswrapper[4679]: E1202 10:20:32.908173 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.908201 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.908448 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:32 crc kubenswrapper[4679]: E1202 10:20:32.908565 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:32 crc kubenswrapper[4679]: E1202 10:20:32.908636 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:32 crc kubenswrapper[4679]: E1202 10:20:32.908718 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.924699 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.993789 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.993816 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.993824 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.993836 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:32 crc kubenswrapper[4679]: I1202 10:20:32.993845 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:32Z","lastTransitionTime":"2025-12-02T10:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.096543 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.096582 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.096593 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.096608 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.096619 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.199236 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.199290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.199317 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.199337 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.199351 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.302718 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.302752 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.302760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.302793 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.302803 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.406172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.406209 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.406218 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.406232 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.406241 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.509086 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.509155 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.509166 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.509184 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.509204 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.611439 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.611476 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.611487 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.611503 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.611514 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.713985 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.714020 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.714032 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.714054 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.714065 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.816256 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.816290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.816298 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.816332 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.816341 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.919181 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.919245 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.919257 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.919272 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:33 crc kubenswrapper[4679]: I1202 10:20:33.919282 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:33Z","lastTransitionTime":"2025-12-02T10:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.021860 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.021899 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.021910 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.021927 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.021939 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.124156 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.124205 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.124214 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.124229 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.124238 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.226653 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.226709 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.226719 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.226737 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.226749 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.328635 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.328668 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.328677 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.328692 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.328703 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.431276 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.431344 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.431355 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.431371 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.431381 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.533685 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.533731 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.533760 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.533776 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.533787 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.636461 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.636536 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.636546 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.636587 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.636600 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.739073 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.739111 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.739120 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.739137 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.739147 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.840881 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.840920 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.840929 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.840944 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.840953 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.907785 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.907835 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.907799 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.907897 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:34 crc kubenswrapper[4679]: E1202 10:20:34.908232 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:34 crc kubenswrapper[4679]: E1202 10:20:34.908375 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:34 crc kubenswrapper[4679]: E1202 10:20:34.908517 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:34 crc kubenswrapper[4679]: E1202 10:20:34.908115 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.943401 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.943449 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.943459 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.943478 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:34 crc kubenswrapper[4679]: I1202 10:20:34.943491 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:34Z","lastTransitionTime":"2025-12-02T10:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.045727 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.045764 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.045775 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.045790 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.045801 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.148289 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.148612 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.148677 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.148739 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.148816 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.252921 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.253202 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.253336 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.253446 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.253545 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.355987 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.356042 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.356054 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.356072 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.356084 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.458143 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.458194 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.458204 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.458226 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.458238 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.560878 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.560921 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.560934 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.560950 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.560961 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.663438 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.663478 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.663486 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.663499 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.663513 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.766226 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.766261 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.766269 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.766281 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.766291 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.868656 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.868690 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.868699 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.868713 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.868722 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.971326 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.971383 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.971399 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.971418 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:35 crc kubenswrapper[4679]: I1202 10:20:35.971430 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:35Z","lastTransitionTime":"2025-12-02T10:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.074472 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.074516 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.074526 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.074544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.074554 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.177169 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.177214 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.177225 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.177243 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.177254 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.280185 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.280433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.280511 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.280653 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.280738 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.382625 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.382899 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.382972 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.383040 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.383104 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.484890 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.485193 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.485290 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.485433 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.485524 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.590071 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.590118 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.590127 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.590145 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.590157 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.692544 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.692585 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.692594 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.692607 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.692615 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.795788 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.795832 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.795847 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.795868 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.795883 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.898001 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.898051 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.898067 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.898089 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.898105 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:36Z","lastTransitionTime":"2025-12-02T10:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.908042 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.908526 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.908607 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.908825 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:36 crc kubenswrapper[4679]: E1202 10:20:36.908986 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:36 crc kubenswrapper[4679]: E1202 10:20:36.909262 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:36 crc kubenswrapper[4679]: E1202 10:20:36.909352 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:36 crc kubenswrapper[4679]: E1202 10:20:36.909419 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.919432 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1827954-0852-40f4-ad94-c979f72addb9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7j2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xlg8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.936393 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e450591-2a0c-4aa3-aad1-073f92727a38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4412d6679752d83da37af46ffca941ecefe450f86c15c9dfe8d6a0208edfbb6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a92f30a6765291b9546ffa6a1a6a72d517a93b555e806474c48662f1695721b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4edb1570dd27a7e1db48be709a42d81ef62e0e322aab9ec8c747e78a272e1143\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.947425 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.956752 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd7e05f6ec5fe6cdbc30ab9d96f425cad9a1dafea8da1628ffebd95efec46fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8lbf6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lzf8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.964573 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7pw64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e687ed45-c025-48d3-9eeb-d4cffad70f50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb69ec5e0ca90c06956f10f5950087dedce49b715279d65bfcc159f49314a76c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzzv7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7pw64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.973878 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5xr66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4b4627f-c062-4e2e-a17b-cb749e9dd917\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0407064373a8bebc0d0415e728edeffc84b6ac3d459475ed16c65e1d42093bb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kgcn5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:26Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5xr66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.986932 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"141dee09-85e8-43f5-a5d5-52458cce375a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5108aadc26e7e021822da496e37be8900d992c2f5faa83313ddc8ece8a50af6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1533da90e1c3da733b39067d2aa3e62a5ad70655c2af96ca5ffac36174269d42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z58h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T10:19:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vs22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:36 crc kubenswrapper[4679]: I1202 10:20:36.996750 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20d830ed6550b218ede322fd978e46b8eed4346be5eb6a15b96ed2dfd0db4246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:36Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:36.999981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.000008 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.000015 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.000028 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.000036 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.011580 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb3b222b9f88df5879abf9e57033ba2bcc9c893939a2dfdef354835159c3afc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01825f6487671dfaa94550aba83a187c58b7ca9a938fbe78549cee24c1a5d776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T10:19:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.023472 4679 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T10:19:24Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T10:20:37Z is after 2025-08-24T17:21:41Z" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.044060 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-r5sft" podStartSLOduration=73.044042271 podStartE2EDuration="1m13.044042271s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:37.043963068 +0000 UTC m=+90.374101928" watchObservedRunningTime="2025-12-02 10:20:37.044042271 +0000 UTC m=+90.374181131" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.084840 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2lmfz" podStartSLOduration=73.084824938 podStartE2EDuration="1m13.084824938s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:37.061771676 +0000 UTC m=+90.391910536" watchObservedRunningTime="2025-12-02 10:20:37.084824938 +0000 UTC m=+90.414963798" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.101922 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.101952 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.101963 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.101977 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.101987 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.114541 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.1145238 podStartE2EDuration="1m11.1145238s" podCreationTimestamp="2025-12-02 10:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:37.112294679 +0000 UTC m=+90.442433559" watchObservedRunningTime="2025-12-02 10:20:37.1145238 +0000 UTC m=+90.444662660" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.131898 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.131879465 podStartE2EDuration="35.131879465s" podCreationTimestamp="2025-12-02 10:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:37.131295229 +0000 UTC m=+90.461434089" watchObservedRunningTime="2025-12-02 10:20:37.131879465 +0000 UTC m=+90.462018325" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.160892 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.160877159 podStartE2EDuration="5.160877159s" podCreationTimestamp="2025-12-02 10:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:37.159967035 +0000 UTC m=+90.490105895" watchObservedRunningTime="2025-12-02 10:20:37.160877159 +0000 UTC m=+90.491016019" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.200467 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.200443783 podStartE2EDuration="28.200443783s" podCreationTimestamp="2025-12-02 10:20:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:37.196739622 +0000 UTC m=+90.526878482" watchObservedRunningTime="2025-12-02 10:20:37.200443783 +0000 UTC m=+90.530582643" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.204103 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.204146 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.204157 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.204172 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.204184 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.306103 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.306150 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.306158 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.306179 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.306193 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.408970 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.409075 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.409098 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.409127 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.409151 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.511743 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.511795 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.511809 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.511828 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.511843 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.614750 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.614913 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.615159 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.615183 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.615588 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.717700 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.717744 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.717787 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.717803 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.717813 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.821020 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.821094 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.821107 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.821145 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.821191 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.923611 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.923646 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.923656 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.923672 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:37 crc kubenswrapper[4679]: I1202 10:20:37.923682 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:37Z","lastTransitionTime":"2025-12-02T10:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.026716 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.026796 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.026840 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.026861 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.026873 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.130260 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.130336 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.130346 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.130361 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.130372 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.233695 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.233800 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.233815 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.233839 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.233854 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.339268 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.339434 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.339475 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.339505 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.339527 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.442360 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.442405 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.442417 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.442434 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.442445 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.544422 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.544464 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.544475 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.544491 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.544503 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.647467 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.647514 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.647524 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.647540 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.647553 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.750708 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.750742 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.750752 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.750765 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.750776 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.854737 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.855161 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.855259 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.855404 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.855503 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.908128 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.908128 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:38 crc kubenswrapper[4679]: E1202 10:20:38.908336 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.908155 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:38 crc kubenswrapper[4679]: E1202 10:20:38.908523 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:38 crc kubenswrapper[4679]: E1202 10:20:38.908648 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.908841 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:38 crc kubenswrapper[4679]: E1202 10:20:38.908983 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.957626 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.957680 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.957700 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.957721 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:38 crc kubenswrapper[4679]: I1202 10:20:38.957738 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:38Z","lastTransitionTime":"2025-12-02T10:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.060561 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.060604 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.060612 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.060627 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.060637 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.162977 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.163045 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.163055 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.163069 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.163078 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.265642 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.265703 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.265715 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.265729 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.265738 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.367816 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.367880 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.367896 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.367919 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.367934 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.470392 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.470427 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.470441 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.470456 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.470465 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.572944 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.572981 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.572991 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.573006 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.573016 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.652265 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.652329 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.652341 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.652357 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.652370 4679 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T10:20:39Z","lastTransitionTime":"2025-12-02T10:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.699914 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc"] Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.700589 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.702755 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.703098 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.704967 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.706612 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.718855 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vs22q" podStartSLOduration=75.718822085 podStartE2EDuration="1m15.718822085s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:39.717232622 +0000 UTC m=+93.047371482" watchObservedRunningTime="2025-12-02 10:20:39.718822085 +0000 UTC m=+93.048960955" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.748327 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=71.748292753 podStartE2EDuration="1m11.748292753s" podCreationTimestamp="2025-12-02 10:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:39.747736557 +0000 UTC m=+93.077875427" watchObservedRunningTime="2025-12-02 10:20:39.748292753 +0000 UTC m=+93.078431613" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.755397 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.755461 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.755507 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.755530 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.755574 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.777481 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podStartSLOduration=75.777464992 podStartE2EDuration="1m15.777464992s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:39.777358909 +0000 UTC m=+93.107497779" watchObservedRunningTime="2025-12-02 10:20:39.777464992 +0000 UTC m=+93.107603852" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.800377 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-7pw64" podStartSLOduration=75.800357688 podStartE2EDuration="1m15.800357688s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:39.788521884 +0000 UTC m=+93.118660744" watchObservedRunningTime="2025-12-02 10:20:39.800357688 +0000 UTC m=+93.130496538" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.814893 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5xr66" podStartSLOduration=75.814875076 podStartE2EDuration="1m15.814875076s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:39.800509953 +0000 UTC m=+93.130648813" watchObservedRunningTime="2025-12-02 10:20:39.814875076 +0000 UTC m=+93.145013936" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.856538 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.856594 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.856623 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.856655 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.856671 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.856940 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.857003 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.857857 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.865286 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:39 crc kubenswrapper[4679]: I1202 10:20:39.872592 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a73278cb-e92d-4fc0-85c0-65c6b0a082ad-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mzxxc\" (UID: \"a73278cb-e92d-4fc0-85c0-65c6b0a082ad\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:40 crc kubenswrapper[4679]: I1202 10:20:40.016611 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" Dec 02 10:20:40 crc kubenswrapper[4679]: W1202 10:20:40.031767 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda73278cb_e92d_4fc0_85c0_65c6b0a082ad.slice/crio-db4f4d23b5803eadd303e5644bee6782040516817579104835fbe380e7114491 WatchSource:0}: Error finding container db4f4d23b5803eadd303e5644bee6782040516817579104835fbe380e7114491: Status 404 returned error can't find the container with id db4f4d23b5803eadd303e5644bee6782040516817579104835fbe380e7114491 Dec 02 10:20:40 crc kubenswrapper[4679]: I1202 10:20:40.364918 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" event={"ID":"a73278cb-e92d-4fc0-85c0-65c6b0a082ad","Type":"ContainerStarted","Data":"db4f4d23b5803eadd303e5644bee6782040516817579104835fbe380e7114491"} Dec 02 10:20:40 crc kubenswrapper[4679]: I1202 10:20:40.908281 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:40 crc kubenswrapper[4679]: I1202 10:20:40.908357 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:40 crc kubenswrapper[4679]: E1202 10:20:40.909077 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:40 crc kubenswrapper[4679]: I1202 10:20:40.908409 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:40 crc kubenswrapper[4679]: I1202 10:20:40.908358 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:40 crc kubenswrapper[4679]: E1202 10:20:40.909353 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:40 crc kubenswrapper[4679]: E1202 10:20:40.909444 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:40 crc kubenswrapper[4679]: E1202 10:20:40.909559 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:41 crc kubenswrapper[4679]: I1202 10:20:41.373474 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" event={"ID":"a73278cb-e92d-4fc0-85c0-65c6b0a082ad","Type":"ContainerStarted","Data":"3e82d3e3e7f4661d54718762bee6ababa8e1488779fbe8e98ad52e1299546c5f"} Dec 02 10:20:41 crc kubenswrapper[4679]: I1202 10:20:41.387499 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzxxc" podStartSLOduration=77.387477755 podStartE2EDuration="1m17.387477755s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:20:41.387401953 +0000 UTC m=+94.717540833" watchObservedRunningTime="2025-12-02 10:20:41.387477755 +0000 UTC m=+94.717616615" Dec 02 10:20:41 crc kubenswrapper[4679]: I1202 10:20:41.908323 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:20:41 crc kubenswrapper[4679]: E1202 10:20:41.908519 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:42 crc kubenswrapper[4679]: I1202 10:20:42.908142 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:42 crc kubenswrapper[4679]: I1202 10:20:42.908190 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:42 crc kubenswrapper[4679]: I1202 10:20:42.908179 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:42 crc kubenswrapper[4679]: I1202 10:20:42.908142 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:42 crc kubenswrapper[4679]: E1202 10:20:42.908290 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:42 crc kubenswrapper[4679]: E1202 10:20:42.908402 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:42 crc kubenswrapper[4679]: E1202 10:20:42.908468 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:42 crc kubenswrapper[4679]: E1202 10:20:42.908532 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:43 crc kubenswrapper[4679]: I1202 10:20:43.095697 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:43 crc kubenswrapper[4679]: E1202 10:20:43.095820 4679 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:20:43 crc kubenswrapper[4679]: E1202 10:20:43.095892 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs podName:c1827954-0852-40f4-ad94-c979f72addb9 nodeName:}" failed. No retries permitted until 2025-12-02 10:21:47.095875835 +0000 UTC m=+160.426014695 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs") pod "network-metrics-daemon-xlg8w" (UID: "c1827954-0852-40f4-ad94-c979f72addb9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 10:20:44 crc kubenswrapper[4679]: I1202 10:20:44.908603 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:44 crc kubenswrapper[4679]: I1202 10:20:44.908665 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:44 crc kubenswrapper[4679]: I1202 10:20:44.908603 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:44 crc kubenswrapper[4679]: E1202 10:20:44.908742 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:44 crc kubenswrapper[4679]: I1202 10:20:44.908613 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:44 crc kubenswrapper[4679]: E1202 10:20:44.908846 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:44 crc kubenswrapper[4679]: E1202 10:20:44.909012 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:44 crc kubenswrapper[4679]: E1202 10:20:44.909056 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:46 crc kubenswrapper[4679]: I1202 10:20:46.908344 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:46 crc kubenswrapper[4679]: I1202 10:20:46.908344 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:46 crc kubenswrapper[4679]: E1202 10:20:46.909946 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:46 crc kubenswrapper[4679]: I1202 10:20:46.909999 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:46 crc kubenswrapper[4679]: I1202 10:20:46.910041 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:46 crc kubenswrapper[4679]: E1202 10:20:46.910109 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:46 crc kubenswrapper[4679]: E1202 10:20:46.910416 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:46 crc kubenswrapper[4679]: E1202 10:20:46.910485 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:48 crc kubenswrapper[4679]: I1202 10:20:48.908445 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:48 crc kubenswrapper[4679]: I1202 10:20:48.908509 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:48 crc kubenswrapper[4679]: E1202 10:20:48.908561 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:48 crc kubenswrapper[4679]: I1202 10:20:48.908606 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:48 crc kubenswrapper[4679]: I1202 10:20:48.908888 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:48 crc kubenswrapper[4679]: E1202 10:20:48.908894 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:48 crc kubenswrapper[4679]: E1202 10:20:48.908954 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:48 crc kubenswrapper[4679]: E1202 10:20:48.909009 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:50 crc kubenswrapper[4679]: I1202 10:20:50.908764 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:50 crc kubenswrapper[4679]: I1202 10:20:50.908872 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:50 crc kubenswrapper[4679]: I1202 10:20:50.908814 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:50 crc kubenswrapper[4679]: I1202 10:20:50.908991 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:50 crc kubenswrapper[4679]: E1202 10:20:50.909284 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:50 crc kubenswrapper[4679]: E1202 10:20:50.909395 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:50 crc kubenswrapper[4679]: E1202 10:20:50.909473 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:50 crc kubenswrapper[4679]: E1202 10:20:50.909612 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:52 crc kubenswrapper[4679]: I1202 10:20:52.908498 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:52 crc kubenswrapper[4679]: I1202 10:20:52.908566 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:52 crc kubenswrapper[4679]: E1202 10:20:52.908641 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:52 crc kubenswrapper[4679]: I1202 10:20:52.908693 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:52 crc kubenswrapper[4679]: I1202 10:20:52.908695 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:52 crc kubenswrapper[4679]: E1202 10:20:52.908788 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:52 crc kubenswrapper[4679]: E1202 10:20:52.909075 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:52 crc kubenswrapper[4679]: E1202 10:20:52.909136 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:54 crc kubenswrapper[4679]: I1202 10:20:54.908159 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:54 crc kubenswrapper[4679]: E1202 10:20:54.908276 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:54 crc kubenswrapper[4679]: I1202 10:20:54.908492 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:54 crc kubenswrapper[4679]: E1202 10:20:54.908546 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:54 crc kubenswrapper[4679]: I1202 10:20:54.908665 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:54 crc kubenswrapper[4679]: I1202 10:20:54.908765 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:54 crc kubenswrapper[4679]: E1202 10:20:54.908875 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:54 crc kubenswrapper[4679]: E1202 10:20:54.908981 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:55 crc kubenswrapper[4679]: I1202 10:20:55.909579 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:20:55 crc kubenswrapper[4679]: E1202 10:20:55.909781 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:20:56 crc kubenswrapper[4679]: I1202 10:20:56.908520 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:56 crc kubenswrapper[4679]: I1202 10:20:56.908537 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:56 crc kubenswrapper[4679]: E1202 10:20:56.910066 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:56 crc kubenswrapper[4679]: I1202 10:20:56.910103 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:56 crc kubenswrapper[4679]: E1202 10:20:56.910903 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:56 crc kubenswrapper[4679]: E1202 10:20:56.910203 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:56 crc kubenswrapper[4679]: I1202 10:20:56.910126 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:56 crc kubenswrapper[4679]: E1202 10:20:56.911289 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.419435 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/1.log" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.420215 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/0.log" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.420261 4679 generic.go:334] "Generic (PLEG): container finished" podID="06f77afa-f710-437b-9cfe-10959bdc1ac8" containerID="97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab" exitCode=1 Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.420301 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerDied","Data":"97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab"} Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.420364 4679 scope.go:117] "RemoveContainer" containerID="e945df8c87ae6743ceba3a424bfd72fb67911e40f796ea1c86190704c08b1876" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.420869 4679 scope.go:117] "RemoveContainer" containerID="97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab" Dec 02 10:20:58 crc kubenswrapper[4679]: E1202 10:20:58.421075 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-r5sft_openshift-multus(06f77afa-f710-437b-9cfe-10959bdc1ac8)\"" pod="openshift-multus/multus-r5sft" podUID="06f77afa-f710-437b-9cfe-10959bdc1ac8" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.908521 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.908543 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.908579 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:20:58 crc kubenswrapper[4679]: I1202 10:20:58.908604 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:20:58 crc kubenswrapper[4679]: E1202 10:20:58.908985 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:20:58 crc kubenswrapper[4679]: E1202 10:20:58.909149 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:20:58 crc kubenswrapper[4679]: E1202 10:20:58.909250 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:20:58 crc kubenswrapper[4679]: E1202 10:20:58.909352 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:20:59 crc kubenswrapper[4679]: I1202 10:20:59.424279 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/1.log" Dec 02 10:21:00 crc kubenswrapper[4679]: I1202 10:21:00.908090 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:00 crc kubenswrapper[4679]: I1202 10:21:00.908142 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:00 crc kubenswrapper[4679]: I1202 10:21:00.908165 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:00 crc kubenswrapper[4679]: E1202 10:21:00.908222 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:00 crc kubenswrapper[4679]: E1202 10:21:00.908296 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:00 crc kubenswrapper[4679]: I1202 10:21:00.908354 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:00 crc kubenswrapper[4679]: E1202 10:21:00.908389 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:00 crc kubenswrapper[4679]: E1202 10:21:00.908477 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:02 crc kubenswrapper[4679]: I1202 10:21:02.908449 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:02 crc kubenswrapper[4679]: E1202 10:21:02.908591 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:02 crc kubenswrapper[4679]: I1202 10:21:02.908662 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:02 crc kubenswrapper[4679]: E1202 10:21:02.908760 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:02 crc kubenswrapper[4679]: I1202 10:21:02.909054 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:02 crc kubenswrapper[4679]: E1202 10:21:02.909114 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:02 crc kubenswrapper[4679]: I1202 10:21:02.909257 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:02 crc kubenswrapper[4679]: E1202 10:21:02.909424 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:04 crc kubenswrapper[4679]: I1202 10:21:04.908301 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:04 crc kubenswrapper[4679]: I1202 10:21:04.908409 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:04 crc kubenswrapper[4679]: I1202 10:21:04.908562 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:04 crc kubenswrapper[4679]: E1202 10:21:04.908551 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:04 crc kubenswrapper[4679]: E1202 10:21:04.908649 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:04 crc kubenswrapper[4679]: E1202 10:21:04.908708 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:04 crc kubenswrapper[4679]: I1202 10:21:04.908377 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:04 crc kubenswrapper[4679]: E1202 10:21:04.909591 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:06 crc kubenswrapper[4679]: I1202 10:21:06.908629 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:06 crc kubenswrapper[4679]: E1202 10:21:06.908726 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:06 crc kubenswrapper[4679]: I1202 10:21:06.908638 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:06 crc kubenswrapper[4679]: I1202 10:21:06.908742 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:06 crc kubenswrapper[4679]: E1202 10:21:06.908936 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:06 crc kubenswrapper[4679]: E1202 10:21:06.908981 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:06 crc kubenswrapper[4679]: I1202 10:21:06.909435 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:06 crc kubenswrapper[4679]: E1202 10:21:06.909528 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:06 crc kubenswrapper[4679]: E1202 10:21:06.919968 4679 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 02 10:21:07 crc kubenswrapper[4679]: E1202 10:21:07.020861 4679 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 10:21:07 crc kubenswrapper[4679]: I1202 10:21:07.909354 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:21:07 crc kubenswrapper[4679]: E1202 10:21:07.909987 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pjb6d_openshift-ovn-kubernetes(a66c9e2c-2ca3-4348-84cc-19f365505c9e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" Dec 02 10:21:08 crc kubenswrapper[4679]: I1202 10:21:08.908507 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:08 crc kubenswrapper[4679]: I1202 10:21:08.908567 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:08 crc kubenswrapper[4679]: E1202 10:21:08.908993 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:08 crc kubenswrapper[4679]: I1202 10:21:08.908703 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:08 crc kubenswrapper[4679]: E1202 10:21:08.909238 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:08 crc kubenswrapper[4679]: I1202 10:21:08.908608 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:08 crc kubenswrapper[4679]: E1202 10:21:08.909998 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:08 crc kubenswrapper[4679]: E1202 10:21:08.908997 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:10 crc kubenswrapper[4679]: I1202 10:21:10.907704 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:10 crc kubenswrapper[4679]: E1202 10:21:10.908543 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:10 crc kubenswrapper[4679]: I1202 10:21:10.907767 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:10 crc kubenswrapper[4679]: I1202 10:21:10.907781 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:10 crc kubenswrapper[4679]: E1202 10:21:10.908697 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:10 crc kubenswrapper[4679]: E1202 10:21:10.908794 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:10 crc kubenswrapper[4679]: I1202 10:21:10.907739 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:10 crc kubenswrapper[4679]: E1202 10:21:10.908920 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:12 crc kubenswrapper[4679]: E1202 10:21:12.022176 4679 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 10:21:12 crc kubenswrapper[4679]: I1202 10:21:12.908112 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:12 crc kubenswrapper[4679]: I1202 10:21:12.908125 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:12 crc kubenswrapper[4679]: E1202 10:21:12.908674 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:12 crc kubenswrapper[4679]: I1202 10:21:12.908346 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:12 crc kubenswrapper[4679]: I1202 10:21:12.908236 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:12 crc kubenswrapper[4679]: E1202 10:21:12.908608 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:12 crc kubenswrapper[4679]: I1202 10:21:12.908457 4679 scope.go:117] "RemoveContainer" containerID="97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab" Dec 02 10:21:12 crc kubenswrapper[4679]: E1202 10:21:12.908778 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:12 crc kubenswrapper[4679]: E1202 10:21:12.908899 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:13 crc kubenswrapper[4679]: I1202 10:21:13.471940 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/1.log" Dec 02 10:21:13 crc kubenswrapper[4679]: I1202 10:21:13.472266 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerStarted","Data":"37cca3a05037016399c4df02f078b7a1a8d43fa20e09c359c1b34e975cdc82f7"} Dec 02 10:21:14 crc kubenswrapper[4679]: I1202 10:21:14.907899 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:14 crc kubenswrapper[4679]: E1202 10:21:14.908030 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:14 crc kubenswrapper[4679]: I1202 10:21:14.908077 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:14 crc kubenswrapper[4679]: I1202 10:21:14.908155 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:14 crc kubenswrapper[4679]: E1202 10:21:14.908238 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:14 crc kubenswrapper[4679]: I1202 10:21:14.908456 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:14 crc kubenswrapper[4679]: E1202 10:21:14.908532 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:14 crc kubenswrapper[4679]: E1202 10:21:14.908621 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:16 crc kubenswrapper[4679]: I1202 10:21:16.908086 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:16 crc kubenswrapper[4679]: I1202 10:21:16.908119 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:16 crc kubenswrapper[4679]: I1202 10:21:16.908082 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:16 crc kubenswrapper[4679]: E1202 10:21:16.909101 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:16 crc kubenswrapper[4679]: I1202 10:21:16.909119 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:16 crc kubenswrapper[4679]: E1202 10:21:16.909243 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:16 crc kubenswrapper[4679]: E1202 10:21:16.909275 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:16 crc kubenswrapper[4679]: E1202 10:21:16.909424 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:17 crc kubenswrapper[4679]: E1202 10:21:17.022776 4679 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 10:21:18 crc kubenswrapper[4679]: I1202 10:21:18.908469 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:18 crc kubenswrapper[4679]: E1202 10:21:18.908678 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:18 crc kubenswrapper[4679]: I1202 10:21:18.909085 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:18 crc kubenswrapper[4679]: E1202 10:21:18.909189 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:18 crc kubenswrapper[4679]: I1202 10:21:18.909473 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:18 crc kubenswrapper[4679]: E1202 10:21:18.909675 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:18 crc kubenswrapper[4679]: I1202 10:21:18.909691 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:18 crc kubenswrapper[4679]: E1202 10:21:18.909915 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:19 crc kubenswrapper[4679]: I1202 10:21:19.908832 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:21:20 crc kubenswrapper[4679]: I1202 10:21:20.908395 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:20 crc kubenswrapper[4679]: I1202 10:21:20.908427 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:20 crc kubenswrapper[4679]: I1202 10:21:20.908542 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:20 crc kubenswrapper[4679]: I1202 10:21:20.908729 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:20 crc kubenswrapper[4679]: E1202 10:21:20.908722 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:20 crc kubenswrapper[4679]: E1202 10:21:20.908821 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:20 crc kubenswrapper[4679]: E1202 10:21:20.908902 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:20 crc kubenswrapper[4679]: E1202 10:21:20.908974 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:21 crc kubenswrapper[4679]: I1202 10:21:21.511201 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/3.log" Dec 02 10:21:21 crc kubenswrapper[4679]: I1202 10:21:21.514282 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerStarted","Data":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} Dec 02 10:21:21 crc kubenswrapper[4679]: I1202 10:21:21.514790 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:21:21 crc kubenswrapper[4679]: I1202 10:21:21.543358 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podStartSLOduration=117.543339082 podStartE2EDuration="1m57.543339082s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:21.542728658 +0000 UTC m=+134.872867518" watchObservedRunningTime="2025-12-02 10:21:21.543339082 +0000 UTC m=+134.873477942" Dec 02 10:21:21 crc kubenswrapper[4679]: I1202 10:21:21.676270 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xlg8w"] Dec 02 10:21:21 crc kubenswrapper[4679]: I1202 10:21:21.676713 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:21 crc kubenswrapper[4679]: E1202 10:21:21.676809 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:22 crc kubenswrapper[4679]: E1202 10:21:22.024075 4679 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 10:21:22 crc kubenswrapper[4679]: I1202 10:21:22.908126 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:22 crc kubenswrapper[4679]: I1202 10:21:22.908126 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:22 crc kubenswrapper[4679]: I1202 10:21:22.908198 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:22 crc kubenswrapper[4679]: E1202 10:21:22.908363 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:22 crc kubenswrapper[4679]: I1202 10:21:22.908440 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:22 crc kubenswrapper[4679]: E1202 10:21:22.908444 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:22 crc kubenswrapper[4679]: E1202 10:21:22.908565 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:22 crc kubenswrapper[4679]: E1202 10:21:22.908717 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:24 crc kubenswrapper[4679]: I1202 10:21:24.908558 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:24 crc kubenswrapper[4679]: I1202 10:21:24.908603 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:24 crc kubenswrapper[4679]: I1202 10:21:24.908693 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:24 crc kubenswrapper[4679]: E1202 10:21:24.908684 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:24 crc kubenswrapper[4679]: I1202 10:21:24.908730 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:24 crc kubenswrapper[4679]: E1202 10:21:24.908857 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:24 crc kubenswrapper[4679]: E1202 10:21:24.908884 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:24 crc kubenswrapper[4679]: E1202 10:21:24.908936 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:26 crc kubenswrapper[4679]: I1202 10:21:26.908549 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:26 crc kubenswrapper[4679]: E1202 10:21:26.911530 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 10:21:26 crc kubenswrapper[4679]: I1202 10:21:26.911735 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:26 crc kubenswrapper[4679]: I1202 10:21:26.911771 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:26 crc kubenswrapper[4679]: I1202 10:21:26.911772 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:26 crc kubenswrapper[4679]: E1202 10:21:26.911814 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 10:21:26 crc kubenswrapper[4679]: E1202 10:21:26.911978 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xlg8w" podUID="c1827954-0852-40f4-ad94-c979f72addb9" Dec 02 10:21:26 crc kubenswrapper[4679]: E1202 10:21:26.912132 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.908054 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.908100 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.908122 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.908348 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.910716 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.912103 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.912131 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.912238 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.912361 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 02 10:21:28 crc kubenswrapper[4679]: I1202 10:21:28.913846 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.693180 4679 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.726781 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s6wzr"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.727385 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.727639 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.727930 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.728542 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.728921 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.730083 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cx5cx"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.730575 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.731348 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.731546 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-z95rb"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.732059 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.732782 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.733199 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.735025 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qcmqt"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.735515 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.735956 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.737056 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.738064 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vnw7n"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.739131 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.741026 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.741664 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.742796 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.742831 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.742981 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.743500 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.743682 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.746544 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.747174 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.747809 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748017 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748040 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748361 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748383 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748538 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748748 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.748875 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.749015 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.749151 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.749338 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.747825 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.750691 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.750936 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.751096 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.751352 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.751494 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.751633 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.751719 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.751892 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.752010 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.752695 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.752861 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753172 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753358 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753438 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753495 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753500 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753612 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753666 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753773 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753878 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.753882 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.754804 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.755323 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.760471 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.760882 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.761286 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.761657 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.762513 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.765120 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.765504 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.765358 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.766405 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.767361 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.892726 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.892743 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.893231 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.893236 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.893420 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.893540 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.893684 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895110 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5bk2f"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895242 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895279 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895442 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895509 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895973 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-zr47g"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.896200 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-7psth"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.896580 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.896671 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.896930 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.897100 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-68gpm"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.895672 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.897483 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.897485 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5nh2\" (UniqueName: \"kubernetes.io/projected/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-kube-api-access-d5nh2\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.897793 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.898115 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.898375 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.898611 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.898621 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.898788 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.899210 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.900480 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.901090 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.901102 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.901788 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s6wzr"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.901818 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.902380 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.902439 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.902520 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.902535 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.902694 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.903519 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cx5cx"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.904588 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.907624 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.910832 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.913145 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.914026 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.923399 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.924343 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.924588 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.924846 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.924947 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.925127 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.925376 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.925567 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.925691 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.925829 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.926158 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.926313 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.926552 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.926750 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.926921 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.927253 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.927767 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.928134 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.928488 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.928762 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.929037 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.929479 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.948145 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.948656 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.974604 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5bk2f"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.974635 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zr47g"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.974646 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vnw7n"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.974654 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qcmqt"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.974663 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-z95rb"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.974671 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.982050 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.982942 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.983001 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.985939 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-68gpm"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.993347 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.993899 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.997104 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.997842 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bf564"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998107 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998160 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-audit\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998177 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8jj25"] Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998190 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-config\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998213 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-oauth-config\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998234 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c99d4c8-d743-4050-bf4b-1710297d1c4f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998252 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3142a842-274b-4193-ac44-19b98f0a3a94-trusted-ca\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998268 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-machine-approver-tls\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998294 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998327 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/143f3a78-2efe-4004-8bc6-fdab837914c1-audit-dir\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998344 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3d1572f5-052f-4d3f-83d4-b631497c1d4f-node-pullsecrets\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998363 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lds5q\" (UniqueName: \"kubernetes.io/projected/3142a842-274b-4193-ac44-19b98f0a3a94-kube-api-access-lds5q\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998384 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7caf2887-b32b-4d3d-9c93-a116076dde2c-serving-cert\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998401 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sl2t\" (UniqueName: \"kubernetes.io/projected/143f3a78-2efe-4004-8bc6-fdab837914c1-kube-api-access-9sl2t\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998420 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-image-import-ca\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998440 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3142a842-274b-4193-ac44-19b98f0a3a94-config\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998456 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/006c66c2-65fc-4cf7-9710-dfc64280b3c0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998473 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/006c66c2-65fc-4cf7-9710-dfc64280b3c0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998493 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998510 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-trusted-ca-bundle\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998546 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998563 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-encryption-config\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998579 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-auth-proxy-config\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998601 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-config\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998622 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-encryption-config\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998643 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ca62df6-c0e6-4872-9302-918bd0c4b39a-serving-cert\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998681 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-client-ca\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998701 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl25v\" (UniqueName: \"kubernetes.io/projected/3d1572f5-052f-4d3f-83d4-b631497c1d4f-kube-api-access-fl25v\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998718 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsgfq\" (UniqueName: \"kubernetes.io/projected/5c99d4c8-d743-4050-bf4b-1710297d1c4f-kube-api-access-bsgfq\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998733 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999021 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.998737 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-audit-policies\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999114 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d1572f5-052f-4d3f-83d4-b631497c1d4f-audit-dir\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999146 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999173 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/006c66c2-65fc-4cf7-9710-dfc64280b3c0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999221 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e222153a-8cab-4ef0-8227-13be178f3f82-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999247 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-config\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999269 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c082626c-da29-486a-988f-c4566ee26c91-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6lmxj\" (UID: \"c082626c-da29-486a-988f-c4566ee26c91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999297 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/7caf2887-b32b-4d3d-9c93-a116076dde2c-kube-api-access-9jzfh\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999339 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-config\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999360 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c99d4c8-d743-4050-bf4b-1710297d1c4f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:30 crc kubenswrapper[4679]: I1202 10:21:30.999382 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m4bq\" (UniqueName: \"kubernetes.io/projected/497dbe86-ba0e-40cb-a0ed-9608281cc6b2-kube-api-access-5m4bq\") pod \"downloads-7954f5f757-zr47g\" (UID: \"497dbe86-ba0e-40cb-a0ed-9608281cc6b2\") " pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999491 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999509 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999577 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p52xk\" (UniqueName: \"kubernetes.io/projected/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-kube-api-access-p52xk\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999625 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8fe15a15-98ee-4b83-98a6-1d09f577aae6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999666 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-config\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999691 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3142a842-274b-4193-ac44-19b98f0a3a94-serving-cert\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999728 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e222153a-8cab-4ef0-8227-13be178f3f82-config\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999755 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999774 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-service-ca-bundle\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999801 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpvdz\" (UniqueName: \"kubernetes.io/projected/5ca62df6-c0e6-4872-9302-918bd0c4b39a-kube-api-access-wpvdz\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999825 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4xxh\" (UniqueName: \"kubernetes.io/projected/c082626c-da29-486a-988f-c4566ee26c91-kube-api-access-l4xxh\") pod \"cluster-samples-operator-665b6dd947-6lmxj\" (UID: \"c082626c-da29-486a-988f-c4566ee26c91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999850 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-serving-cert\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999870 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwkw\" (UniqueName: \"kubernetes.io/projected/e222153a-8cab-4ef0-8227-13be178f3f82-kube-api-access-dwwkw\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999911 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-etcd-client\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999935 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfpt6\" (UniqueName: \"kubernetes.io/projected/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-kube-api-access-kfpt6\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999970 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-serving-cert\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:30.999995 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5mcz\" (UniqueName: \"kubernetes.io/projected/9821b1bc-27fb-4720-afce-16072b779105-kube-api-access-d5mcz\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000015 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-service-ca\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000039 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-oauth-serving-cert\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000061 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q5h4\" (UniqueName: \"kubernetes.io/projected/006c66c2-65fc-4cf7-9710-dfc64280b3c0-kube-api-access-2q5h4\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000111 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5nh2\" (UniqueName: \"kubernetes.io/projected/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-kube-api-access-d5nh2\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000151 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7nll\" (UniqueName: \"kubernetes.io/projected/8fe15a15-98ee-4b83-98a6-1d09f577aae6-kube-api-access-l7nll\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000174 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-etcd-serving-ca\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000203 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e222153a-8cab-4ef0-8227-13be178f3f82-images\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000451 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-serving-cert\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000513 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-client-ca\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000551 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-etcd-client\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000576 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-config\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000612 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe15a15-98ee-4b83-98a6-1d09f577aae6-serving-cert\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000636 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9821b1bc-27fb-4720-afce-16072b779105-serving-cert\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.000688 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.001086 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.001248 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.003434 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.004453 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.008004 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h2ljp"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.012472 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.012937 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.013327 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.013948 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.014493 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.016072 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-b689q"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.028382 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.029869 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vmxn8"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.030116 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.030557 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.031604 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.031731 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.043423 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.044129 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.044586 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.045639 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.046041 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.046832 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.047778 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.050033 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.054235 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.065735 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.074669 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.076016 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.076285 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.082731 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.084436 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.085234 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.085779 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.098703 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.100554 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.100562 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.100797 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110541 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-etcd-client\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110585 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfpt6\" (UniqueName: \"kubernetes.io/projected/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-kube-api-access-kfpt6\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110614 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-dir\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110634 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-serving-cert\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110653 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5mcz\" (UniqueName: \"kubernetes.io/projected/9821b1bc-27fb-4720-afce-16072b779105-kube-api-access-d5mcz\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110673 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-service-ca\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110691 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-oauth-serving-cert\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110712 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q5h4\" (UniqueName: \"kubernetes.io/projected/006c66c2-65fc-4cf7-9710-dfc64280b3c0-kube-api-access-2q5h4\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110732 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110764 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7nll\" (UniqueName: \"kubernetes.io/projected/8fe15a15-98ee-4b83-98a6-1d09f577aae6-kube-api-access-l7nll\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110786 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-etcd-serving-ca\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110825 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e222153a-8cab-4ef0-8227-13be178f3f82-images\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110846 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-serving-cert\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110869 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-client-ca\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110891 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110919 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-etcd-client\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110941 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-config\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110964 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.110986 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111011 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe15a15-98ee-4b83-98a6-1d09f577aae6-serving-cert\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111034 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9821b1bc-27fb-4720-afce-16072b779105-serving-cert\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111055 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111077 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-audit\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111124 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111152 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq5fb\" (UniqueName: \"kubernetes.io/projected/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-kube-api-access-dq5fb\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111182 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-config\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111206 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-oauth-config\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111231 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c99d4c8-d743-4050-bf4b-1710297d1c4f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111254 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-policies\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111276 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3142a842-274b-4193-ac44-19b98f0a3a94-trusted-ca\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111296 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-machine-approver-tls\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111353 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111375 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/143f3a78-2efe-4004-8bc6-fdab837914c1-audit-dir\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111396 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3d1572f5-052f-4d3f-83d4-b631497c1d4f-node-pullsecrets\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111472 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lds5q\" (UniqueName: \"kubernetes.io/projected/3142a842-274b-4193-ac44-19b98f0a3a94-kube-api-access-lds5q\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111508 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111533 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df7188a-a756-44c4-988e-81fd270db7aa-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111561 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7caf2887-b32b-4d3d-9c93-a116076dde2c-serving-cert\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111625 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92w6k\" (UniqueName: \"kubernetes.io/projected/2df7188a-a756-44c4-988e-81fd270db7aa-kube-api-access-92w6k\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111752 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sl2t\" (UniqueName: \"kubernetes.io/projected/143f3a78-2efe-4004-8bc6-fdab837914c1-kube-api-access-9sl2t\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111816 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-image-import-ca\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111841 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3142a842-274b-4193-ac44-19b98f0a3a94-config\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111864 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/006c66c2-65fc-4cf7-9710-dfc64280b3c0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111952 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/006c66c2-65fc-4cf7-9710-dfc64280b3c0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.111990 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-trusted-ca-bundle\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112007 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112026 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-encryption-config\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112042 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-auth-proxy-config\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112060 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-config\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112077 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-encryption-config\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112093 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ca62df6-c0e6-4872-9302-918bd0c4b39a-serving-cert\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112124 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-client-ca\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112140 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl25v\" (UniqueName: \"kubernetes.io/projected/3d1572f5-052f-4d3f-83d4-b631497c1d4f-kube-api-access-fl25v\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsgfq\" (UniqueName: \"kubernetes.io/projected/5c99d4c8-d743-4050-bf4b-1710297d1c4f-kube-api-access-bsgfq\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112176 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112196 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-audit-policies\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112213 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d1572f5-052f-4d3f-83d4-b631497c1d4f-audit-dir\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112230 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112249 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/006c66c2-65fc-4cf7-9710-dfc64280b3c0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112275 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e222153a-8cab-4ef0-8227-13be178f3f82-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112335 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-config\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112371 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c082626c-da29-486a-988f-c4566ee26c91-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6lmxj\" (UID: \"c082626c-da29-486a-988f-c4566ee26c91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112388 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112409 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2df7188a-a756-44c4-988e-81fd270db7aa-metrics-tls\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112428 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/7caf2887-b32b-4d3d-9c93-a116076dde2c-kube-api-access-9jzfh\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112445 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-config\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112460 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c99d4c8-d743-4050-bf4b-1710297d1c4f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112476 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m4bq\" (UniqueName: \"kubernetes.io/projected/497dbe86-ba0e-40cb-a0ed-9608281cc6b2-kube-api-access-5m4bq\") pod \"downloads-7954f5f757-zr47g\" (UID: \"497dbe86-ba0e-40cb-a0ed-9608281cc6b2\") " pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112500 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112516 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p52xk\" (UniqueName: \"kubernetes.io/projected/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-kube-api-access-p52xk\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112532 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112547 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df7188a-a756-44c4-988e-81fd270db7aa-trusted-ca\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112565 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8fe15a15-98ee-4b83-98a6-1d09f577aae6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112582 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-config\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112597 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3142a842-274b-4193-ac44-19b98f0a3a94-serving-cert\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112647 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e222153a-8cab-4ef0-8227-13be178f3f82-config\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112663 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112678 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-service-ca-bundle\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112694 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpvdz\" (UniqueName: \"kubernetes.io/projected/5ca62df6-c0e6-4872-9302-918bd0c4b39a-kube-api-access-wpvdz\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112715 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112732 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4xxh\" (UniqueName: \"kubernetes.io/projected/c082626c-da29-486a-988f-c4566ee26c91-kube-api-access-l4xxh\") pod \"cluster-samples-operator-665b6dd947-6lmxj\" (UID: \"c082626c-da29-486a-988f-c4566ee26c91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112748 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-serving-cert\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112769 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwkw\" (UniqueName: \"kubernetes.io/projected/e222153a-8cab-4ef0-8227-13be178f3f82-kube-api-access-dwwkw\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.114987 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.115148 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-etcd-client\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.116109 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.116659 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.116912 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-764ws"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.117114 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe15a15-98ee-4b83-98a6-1d09f577aae6-serving-cert\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.117270 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-client-ca\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.117562 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-serving-cert\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.118015 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-audit-policies\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.118058 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-etcd-serving-ca\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.118082 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d1572f5-052f-4d3f-83d4-b631497c1d4f-audit-dir\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.112205 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-config\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.118731 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e222153a-8cab-4ef0-8227-13be178f3f82-images\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.118983 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h55w5"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.119239 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-oauth-serving-cert\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.119348 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.119526 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.119737 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8fe15a15-98ee-4b83-98a6-1d09f577aae6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.119847 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.120094 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9821b1bc-27fb-4720-afce-16072b779105-serving-cert\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.120577 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-config\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.120734 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.120854 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-audit\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.121938 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-config\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.122153 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-serving-cert\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.122830 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-client-ca\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.123225 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/006c66c2-65fc-4cf7-9710-dfc64280b3c0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.123471 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-service-ca\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.123714 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hmbk"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.124213 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-config\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.124459 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3142a842-274b-4193-ac44-19b98f0a3a94-serving-cert\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.124731 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.124984 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e222153a-8cab-4ef0-8227-13be178f3f82-config\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.125177 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-zl899"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.125673 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.125863 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-etcd-client\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.126026 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.126438 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7caf2887-b32b-4d3d-9c93-a116076dde2c-serving-cert\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.126733 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.126906 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/143f3a78-2efe-4004-8bc6-fdab837914c1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.127521 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/006c66c2-65fc-4cf7-9710-dfc64280b3c0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.127585 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c99d4c8-d743-4050-bf4b-1710297d1c4f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.127819 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3142a842-274b-4193-ac44-19b98f0a3a94-trusted-ca\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.128499 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-trusted-ca-bundle\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.128517 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3142a842-274b-4193-ac44-19b98f0a3a94-config\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.128567 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3d1572f5-052f-4d3f-83d4-b631497c1d4f-node-pullsecrets\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.128622 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/143f3a78-2efe-4004-8bc6-fdab837914c1-audit-dir\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.129139 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-image-import-ca\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.129263 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bbxkc"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130016 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e222153a-8cab-4ef0-8227-13be178f3f82-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130139 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-oauth-config\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130243 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c99d4c8-d743-4050-bf4b-1710297d1c4f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130408 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kvtft"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130751 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-auth-proxy-config\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130820 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130852 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130877 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.130878 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-config\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.131537 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-service-ca-bundle\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.131703 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d1572f5-052f-4d3f-83d4-b631497c1d4f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.132354 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-config\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.132450 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c082626c-da29-486a-988f-c4566ee26c91-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6lmxj\" (UID: \"c082626c-da29-486a-988f-c4566ee26c91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.132914 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d1572f5-052f-4d3f-83d4-b631497c1d4f-encryption-config\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.133003 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.133272 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.133426 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/143f3a78-2efe-4004-8bc6-fdab837914c1-encryption-config\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.133540 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ca62df6-c0e6-4872-9302-918bd0c4b39a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.133826 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-serving-cert\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.134723 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.136029 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ca62df6-c0e6-4872-9302-918bd0c4b39a-serving-cert\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.137479 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.138228 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-machine-approver-tls\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.138899 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.140334 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vmxn8"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.141763 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h2ljp"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.143176 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hmbk"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.144586 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-b689q"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.145971 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-764ws"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.147355 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h55w5"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.150483 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.150703 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kvtft"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.151163 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.152343 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.153629 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-x9sld"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.154742 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.156175 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.157295 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.158679 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.160375 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8jj25"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.161611 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bbxkc"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.163221 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.164069 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.165520 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.167163 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.168534 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x9sld"] Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.170526 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.190899 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.210409 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213246 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213283 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2df7188a-a756-44c4-988e-81fd270db7aa-metrics-tls\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213360 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213649 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df7188a-a756-44c4-988e-81fd270db7aa-trusted-ca\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213695 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213734 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-dir\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213776 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213826 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213857 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213881 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213906 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213936 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq5fb\" (UniqueName: \"kubernetes.io/projected/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-kube-api-access-dq5fb\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.213963 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-policies\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.214152 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.214587 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-dir\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.215261 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.215633 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-policies\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.215724 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df7188a-a756-44c4-988e-81fd270db7aa-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.215831 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92w6k\" (UniqueName: \"kubernetes.io/projected/2df7188a-a756-44c4-988e-81fd270db7aa-kube-api-access-92w6k\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.215894 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.215956 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.216267 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.216400 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df7188a-a756-44c4-988e-81fd270db7aa-trusted-ca\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.216947 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.218441 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.224678 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.224969 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2df7188a-a756-44c4-988e-81fd270db7aa-metrics-tls\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.224987 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.225464 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.226706 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.226840 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.228593 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.229023 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.230059 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.250727 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.286117 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5nh2\" (UniqueName: \"kubernetes.io/projected/933d15a8-efaf-4ff8-97d3-7a0430af5bfc-kube-api-access-d5nh2\") pod \"openshift-apiserver-operator-796bbdcf4f-wppmh\" (UID: \"933d15a8-efaf-4ff8-97d3-7a0430af5bfc\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.313239 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.330639 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.351124 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.351275 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.371175 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.389976 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.410599 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.441465 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.450534 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.470888 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.492016 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.511160 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.520692 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh"] Dec 02 10:21:31 crc kubenswrapper[4679]: W1202 10:21:31.528850 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod933d15a8_efaf_4ff8_97d3_7a0430af5bfc.slice/crio-6fb60a775d6bf4c0d8c088b4dfd93a6ed053893635a9ba4225417462f69368b0 WatchSource:0}: Error finding container 6fb60a775d6bf4c0d8c088b4dfd93a6ed053893635a9ba4225417462f69368b0: Status 404 returned error can't find the container with id 6fb60a775d6bf4c0d8c088b4dfd93a6ed053893635a9ba4225417462f69368b0 Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.531416 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.546450 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" event={"ID":"933d15a8-efaf-4ff8-97d3-7a0430af5bfc","Type":"ContainerStarted","Data":"6fb60a775d6bf4c0d8c088b4dfd93a6ed053893635a9ba4225417462f69368b0"} Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.550227 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.571073 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.591060 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.611834 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.631012 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.650696 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.670825 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.690897 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.711950 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.732132 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.750986 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.770333 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.790953 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.811150 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.831372 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.851362 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.871071 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.890248 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.909981 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.931078 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.950758 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.971791 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 02 10:21:31 crc kubenswrapper[4679]: I1202 10:21:31.990806 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.010827 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.030585 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.051206 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.070773 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.089366 4679 request.go:700] Waited for 1.012652954s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-controller-dockercfg-c2lfx&limit=500&resourceVersion=0 Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.091169 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.111540 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.131278 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.150648 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.171648 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.190669 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.210480 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.250342 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.290211 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwkw\" (UniqueName: \"kubernetes.io/projected/e222153a-8cab-4ef0-8227-13be178f3f82-kube-api-access-dwwkw\") pod \"machine-api-operator-5694c8668f-z95rb\" (UID: \"e222153a-8cab-4ef0-8227-13be178f3f82\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.306835 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfpt6\" (UniqueName: \"kubernetes.io/projected/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-kube-api-access-kfpt6\") pod \"console-f9d7485db-5bk2f\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.311621 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.318112 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.346857 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7nll\" (UniqueName: \"kubernetes.io/projected/8fe15a15-98ee-4b83-98a6-1d09f577aae6-kube-api-access-l7nll\") pod \"openshift-config-operator-7777fb866f-wnwnx\" (UID: \"8fe15a15-98ee-4b83-98a6-1d09f577aae6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.365088 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5mcz\" (UniqueName: \"kubernetes.io/projected/9821b1bc-27fb-4720-afce-16072b779105-kube-api-access-d5mcz\") pod \"controller-manager-879f6c89f-cx5cx\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.384203 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl25v\" (UniqueName: \"kubernetes.io/projected/3d1572f5-052f-4d3f-83d4-b631497c1d4f-kube-api-access-fl25v\") pod \"apiserver-76f77b778f-s6wzr\" (UID: \"3d1572f5-052f-4d3f-83d4-b631497c1d4f\") " pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.401205 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.412001 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsgfq\" (UniqueName: \"kubernetes.io/projected/5c99d4c8-d743-4050-bf4b-1710297d1c4f-kube-api-access-bsgfq\") pod \"openshift-controller-manager-operator-756b6f6bc6-lsz66\" (UID: \"5c99d4c8-d743-4050-bf4b-1710297d1c4f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.424150 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.427613 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q5h4\" (UniqueName: \"kubernetes.io/projected/006c66c2-65fc-4cf7-9710-dfc64280b3c0-kube-api-access-2q5h4\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.445292 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.450094 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p52xk\" (UniqueName: \"kubernetes.io/projected/d91d02e1-27b2-41c4-aedd-8199c2f6eb42-kube-api-access-p52xk\") pod \"machine-approver-56656f9798-7psth\" (UID: \"d91d02e1-27b2-41c4-aedd-8199c2f6eb42\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.456478 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.471160 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.491098 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.506807 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5bk2f"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.510344 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: W1202 10:21:32.527165 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d97c73c_99b7_452c_9e31_f7c99d4c93f3.slice/crio-b99003e690139e98d2c252d58c2dde8dfbc7bf76b57f3438f9da61bce9a1a90a WatchSource:0}: Error finding container b99003e690139e98d2c252d58c2dde8dfbc7bf76b57f3438f9da61bce9a1a90a: Status 404 returned error can't find the container with id b99003e690139e98d2c252d58c2dde8dfbc7bf76b57f3438f9da61bce9a1a90a Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.531055 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.550475 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.551077 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bk2f" event={"ID":"9d97c73c-99b7-452c-9e31-f7c99d4c93f3","Type":"ContainerStarted","Data":"b99003e690139e98d2c252d58c2dde8dfbc7bf76b57f3438f9da61bce9a1a90a"} Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.552626 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.553121 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" event={"ID":"933d15a8-efaf-4ff8-97d3-7a0430af5bfc","Type":"ContainerStarted","Data":"b1eef5a2cbcb09beeaa33e0c675b1511c8da2a6477be3a4a0db8f560ad28b89f"} Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.570504 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.595599 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.611484 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.650655 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s6wzr"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.653746 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.654343 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/7caf2887-b32b-4d3d-9c93-a116076dde2c-kube-api-access-9jzfh\") pod \"route-controller-manager-6576b87f9c-p49dd\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.659585 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" Dec 02 10:21:32 crc kubenswrapper[4679]: W1202 10:21:32.660003 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d1572f5_052f_4d3f_83d4_b631497c1d4f.slice/crio-857eeebcacca766395b9a9011f5e28da6468e6d92e1c2e82f12a688d45b75e24 WatchSource:0}: Error finding container 857eeebcacca766395b9a9011f5e28da6468e6d92e1c2e82f12a688d45b75e24: Status 404 returned error can't find the container with id 857eeebcacca766395b9a9011f5e28da6468e6d92e1c2e82f12a688d45b75e24 Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.671236 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.686737 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.692655 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.693825 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.705427 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cx5cx"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.711641 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.750703 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.753514 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m4bq\" (UniqueName: \"kubernetes.io/projected/497dbe86-ba0e-40cb-a0ed-9608281cc6b2-kube-api-access-5m4bq\") pod \"downloads-7954f5f757-zr47g\" (UID: \"497dbe86-ba0e-40cb-a0ed-9608281cc6b2\") " pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.761511 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-z95rb"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.771590 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.791166 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.799258 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.810668 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.843049 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.843788 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.843900 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:32 crc kubenswrapper[4679]: E1202 10:21:32.844466 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:23:34.844430927 +0000 UTC m=+268.174569787 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.847826 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/006c66c2-65fc-4cf7-9710-dfc64280b3c0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-nrwbn\" (UID: \"006c66c2-65fc-4cf7-9710-dfc64280b3c0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.847856 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.856986 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.859268 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.870990 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.900521 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.908684 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lds5q\" (UniqueName: \"kubernetes.io/projected/3142a842-274b-4193-ac44-19b98f0a3a94-kube-api-access-lds5q\") pod \"console-operator-58897d9998-vnw7n\" (UID: \"3142a842-274b-4193-ac44-19b98f0a3a94\") " pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.926631 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sl2t\" (UniqueName: \"kubernetes.io/projected/143f3a78-2efe-4004-8bc6-fdab837914c1-kube-api-access-9sl2t\") pod \"apiserver-7bbb656c7d-cjrnm\" (UID: \"143f3a78-2efe-4004-8bc6-fdab837914c1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.930857 4679 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.943232 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.950819 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: W1202 10:21:32.957618 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7caf2887_b32b_4d3d_9c93_a116076dde2c.slice/crio-ba18f9ff04c3e78332adcbe93af0667c17e6f0b70d95597e341599be4072bce0 WatchSource:0}: Error finding container ba18f9ff04c3e78332adcbe93af0667c17e6f0b70d95597e341599be4072bce0: Status 404 returned error can't find the container with id ba18f9ff04c3e78332adcbe93af0667c17e6f0b70d95597e341599be4072bce0 Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.969972 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.972448 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66"] Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.973362 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 02 10:21:32 crc kubenswrapper[4679]: I1202 10:21:32.993222 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 02 10:21:32 crc kubenswrapper[4679]: W1202 10:21:32.994569 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c99d4c8_d743_4050_bf4b_1710297d1c4f.slice/crio-38b4a59349a8fa5f1c9096192075ea91cdea6f2dc88a2c2ad675af251b296702 WatchSource:0}: Error finding container 38b4a59349a8fa5f1c9096192075ea91cdea6f2dc88a2c2ad675af251b296702: Status 404 returned error can't find the container with id 38b4a59349a8fa5f1c9096192075ea91cdea6f2dc88a2c2ad675af251b296702 Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.010452 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.033747 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.046706 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.047237 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.057239 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.057504 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.057613 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.083148 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.089421 4679 request.go:700] Waited for 1.958223616s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.091061 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.093006 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpvdz\" (UniqueName: \"kubernetes.io/projected/5ca62df6-c0e6-4872-9302-918bd0c4b39a-kube-api-access-wpvdz\") pod \"authentication-operator-69f744f599-qcmqt\" (UID: \"5ca62df6-c0e6-4872-9302-918bd0c4b39a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.115168 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.124918 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.130199 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.130420 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4xxh\" (UniqueName: \"kubernetes.io/projected/c082626c-da29-486a-988f-c4566ee26c91-kube-api-access-l4xxh\") pod \"cluster-samples-operator-665b6dd947-6lmxj\" (UID: \"c082626c-da29-486a-988f-c4566ee26c91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.133092 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.139876 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.140001 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.150599 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.172397 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.177639 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.208244 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq5fb\" (UniqueName: \"kubernetes.io/projected/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-kube-api-access-dq5fb\") pod \"oauth-openshift-558db77b4-68gpm\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.212279 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.230566 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df7188a-a756-44c4-988e-81fd270db7aa-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.250766 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92w6k\" (UniqueName: \"kubernetes.io/projected/2df7188a-a756-44c4-988e-81fd270db7aa-kube-api-access-92w6k\") pod \"ingress-operator-5b745b69d9-xjmb2\" (UID: \"2df7188a-a756-44c4-988e-81fd270db7aa\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.272167 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zr47g"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.280501 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.297722 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358189 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2af6d859-3308-4f34-a2b8-cfb2f697bd46-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-b689q\" (UID: \"2af6d859-3308-4f34-a2b8-cfb2f697bd46\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358231 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-config\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358261 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-proxy-tls\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358275 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a9ac2-d37f-4b13-bdf1-9705195350be-config\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358291 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/472551c7-c69c-4215-a7e8-22b9fc846cad-srv-cert\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358559 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlw27\" (UniqueName: \"kubernetes.io/projected/95274f18-9c03-4d32-b14f-621646f64534-kube-api-access-jlw27\") pod \"migrator-59844c95c7-lzh87\" (UID: \"95274f18-9c03-4d32-b14f-621646f64534\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358582 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-default-certificate\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358597 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-proxy-tls\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358620 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358638 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38053e83-e26a-42e0-8a7c-7ebfd0714785-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358653 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnkzw\" (UniqueName: \"kubernetes.io/projected/472551c7-c69c-4215-a7e8-22b9fc846cad-kube-api-access-mnkzw\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358668 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-stats-auth\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358686 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88csv\" (UniqueName: \"kubernetes.io/projected/94b03a02-dc7e-43ba-8183-934fb6649cbf-kube-api-access-88csv\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358704 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhqh4\" (UniqueName: \"kubernetes.io/projected/0f5b6f36-f737-42b9-9403-db45098c682a-kube-api-access-jhqh4\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358728 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-964qh\" (UniqueName: \"kubernetes.io/projected/2af6d859-3308-4f34-a2b8-cfb2f697bd46-kube-api-access-964qh\") pod \"multus-admission-controller-857f4d67dd-b689q\" (UID: \"2af6d859-3308-4f34-a2b8-cfb2f697bd46\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358752 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-images\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358766 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94b03a02-dc7e-43ba-8183-934fb6649cbf-srv-cert\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358779 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/679a9ac2-d37f-4b13-bdf1-9705195350be-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358815 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358856 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/877a5c7e-5f1d-4404-b7ec-e8df15385dee-config\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358879 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqkh2\" (UniqueName: \"kubernetes.io/projected/6e31c209-d885-4235-87a7-972b045a6c72-kube-api-access-tqkh2\") pod \"control-plane-machine-set-operator-78cbb6b69f-ffc27\" (UID: \"6e31c209-d885-4235-87a7-972b045a6c72\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358901 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358934 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/679a9ac2-d37f-4b13-bdf1-9705195350be-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358954 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38053e83-e26a-42e0-8a7c-7ebfd0714785-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358974 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-client\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.358992 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-metrics-certs\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359022 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c942617e-6e69-42e7-aa2b-62e634b14583-metrics-tls\") pod \"dns-operator-744455d44c-8jj25\" (UID: \"c942617e-6e69-42e7-aa2b-62e634b14583\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359049 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8ggt\" (UniqueName: \"kubernetes.io/projected/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-kube-api-access-q8ggt\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359133 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359167 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4743c41d-8037-4e64-bc07-befafeb23482-tmpfs\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359192 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-serving-cert\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359228 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38053e83-e26a-42e0-8a7c-7ebfd0714785-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359250 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/472551c7-c69c-4215-a7e8-22b9fc846cad-profile-collector-cert\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359272 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-trusted-ca\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.359288 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-ca\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.360490 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4743c41d-8037-4e64-bc07-befafeb23482-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.360592 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.360754 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-certificates\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.362933 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m9xx\" (UniqueName: \"kubernetes.io/projected/c942617e-6e69-42e7-aa2b-62e634b14583-kube-api-access-8m9xx\") pod \"dns-operator-744455d44c-8jj25\" (UID: \"c942617e-6e69-42e7-aa2b-62e634b14583\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.362970 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnwp4\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-kube-api-access-xnwp4\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.362988 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/877a5c7e-5f1d-4404-b7ec-e8df15385dee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363009 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363025 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94b03a02-dc7e-43ba-8183-934fb6649cbf-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363054 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/877a5c7e-5f1d-4404-b7ec-e8df15385dee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363095 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363244 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltrlw\" (UniqueName: \"kubernetes.io/projected/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-kube-api-access-ltrlw\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363262 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f5b6f36-f737-42b9-9403-db45098c682a-service-ca-bundle\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363279 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-bound-sa-token\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363296 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zh5z\" (UniqueName: \"kubernetes.io/projected/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-kube-api-access-6zh5z\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363330 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-service-ca\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363413 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4743c41d-8037-4e64-bc07-befafeb23482-webhook-cert\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363449 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-tls\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363464 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k96fw\" (UniqueName: \"kubernetes.io/projected/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-kube-api-access-k96fw\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363499 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmmpk\" (UniqueName: \"kubernetes.io/projected/4743c41d-8037-4e64-bc07-befafeb23482-kube-api-access-vmmpk\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.363515 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6e31c209-d885-4235-87a7-972b045a6c72-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ffc27\" (UID: \"6e31c209-d885-4235-87a7-972b045a6c72\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.364768 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:33.864755696 +0000 UTC m=+147.194894556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.379610 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.465394 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.465731 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:33.965637518 +0000 UTC m=+147.295776388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.465984 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d40d1aca-74a1-44e7-aed1-c595ce552803-secret-volume\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466053 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/682f968a-36d2-4da6-bc70-8cbb1eef8b8d-cert\") pod \"ingress-canary-764ws\" (UID: \"682f968a-36d2-4da6-bc70-8cbb1eef8b8d\") " pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466079 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4743c41d-8037-4e64-bc07-befafeb23482-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466105 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466140 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-certificates\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466164 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m9xx\" (UniqueName: \"kubernetes.io/projected/c942617e-6e69-42e7-aa2b-62e634b14583-kube-api-access-8m9xx\") pod \"dns-operator-744455d44c-8jj25\" (UID: \"c942617e-6e69-42e7-aa2b-62e634b14583\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466193 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chvjg\" (UniqueName: \"kubernetes.io/projected/9621be61-f243-4e39-8b36-e7f8ccc3206d-kube-api-access-chvjg\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466217 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnwp4\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-kube-api-access-xnwp4\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466239 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/877a5c7e-5f1d-4404-b7ec-e8df15385dee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466276 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466297 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94b03a02-dc7e-43ba-8183-934fb6649cbf-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466337 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/877a5c7e-5f1d-4404-b7ec-e8df15385dee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466356 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d40d1aca-74a1-44e7-aed1-c595ce552803-config-volume\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466394 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e4fa17b-058b-46c2-8383-fc1fd5ae9b56-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9fzfq\" (UID: \"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466421 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466441 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e37f501-ca1f-43d6-aace-f2a716cb5d79-config-volume\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466463 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq5k8\" (UniqueName: \"kubernetes.io/projected/d40d1aca-74a1-44e7-aed1-c595ce552803-kube-api-access-kq5k8\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466491 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltrlw\" (UniqueName: \"kubernetes.io/projected/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-kube-api-access-ltrlw\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466513 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f5b6f36-f737-42b9-9403-db45098c682a-service-ca-bundle\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466533 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-bound-sa-token\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466556 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zh5z\" (UniqueName: \"kubernetes.io/projected/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-kube-api-access-6zh5z\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466595 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-service-ca\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466626 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466658 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5jr\" (UniqueName: \"kubernetes.io/projected/d079b995-5285-4810-912b-0f41cde4a9d9-kube-api-access-vl5jr\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466687 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-socket-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466716 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4743c41d-8037-4e64-bc07-befafeb23482-webhook-cert\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466743 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hstng\" (UniqueName: \"kubernetes.io/projected/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-kube-api-access-hstng\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466770 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-tls\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466791 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-signing-cabundle\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466812 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-plugins-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466833 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k96fw\" (UniqueName: \"kubernetes.io/projected/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-kube-api-access-k96fw\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466856 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz5hf\" (UniqueName: \"kubernetes.io/projected/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-kube-api-access-wz5hf\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466880 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/9e4fa17b-058b-46c2-8383-fc1fd5ae9b56-kube-api-access-c4qfb\") pod \"package-server-manager-789f6589d5-9fzfq\" (UID: \"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466904 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-certs\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466930 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmmpk\" (UniqueName: \"kubernetes.io/projected/4743c41d-8037-4e64-bc07-befafeb23482-kube-api-access-vmmpk\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.466976 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6e31c209-d885-4235-87a7-972b045a6c72-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ffc27\" (UID: \"6e31c209-d885-4235-87a7-972b045a6c72\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467000 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-registration-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467027 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2af6d859-3308-4f34-a2b8-cfb2f697bd46-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-b689q\" (UID: \"2af6d859-3308-4f34-a2b8-cfb2f697bd46\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467049 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-config\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467072 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-mountpoint-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467096 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhb88\" (UniqueName: \"kubernetes.io/projected/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-kube-api-access-nhb88\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467122 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-proxy-tls\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467147 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a9ac2-d37f-4b13-bdf1-9705195350be-config\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467169 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/472551c7-c69c-4215-a7e8-22b9fc846cad-srv-cert\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467193 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lddl\" (UniqueName: \"kubernetes.io/projected/1e37f501-ca1f-43d6-aace-f2a716cb5d79-kube-api-access-2lddl\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467216 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-signing-key\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467241 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlw27\" (UniqueName: \"kubernetes.io/projected/95274f18-9c03-4d32-b14f-621646f64534-kube-api-access-jlw27\") pod \"migrator-59844c95c7-lzh87\" (UID: \"95274f18-9c03-4d32-b14f-621646f64534\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467264 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-default-certificate\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467290 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-proxy-tls\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467334 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38053e83-e26a-42e0-8a7c-7ebfd0714785-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467361 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnkzw\" (UniqueName: \"kubernetes.io/projected/472551c7-c69c-4215-a7e8-22b9fc846cad-kube-api-access-mnkzw\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467382 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-stats-auth\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467404 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467425 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88csv\" (UniqueName: \"kubernetes.io/projected/94b03a02-dc7e-43ba-8183-934fb6649cbf-kube-api-access-88csv\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467447 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhqh4\" (UniqueName: \"kubernetes.io/projected/0f5b6f36-f737-42b9-9403-db45098c682a-kube-api-access-jhqh4\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467474 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-964qh\" (UniqueName: \"kubernetes.io/projected/2af6d859-3308-4f34-a2b8-cfb2f697bd46-kube-api-access-964qh\") pod \"multus-admission-controller-857f4d67dd-b689q\" (UID: \"2af6d859-3308-4f34-a2b8-cfb2f697bd46\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467497 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-images\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467521 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94b03a02-dc7e-43ba-8183-934fb6649cbf-srv-cert\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467544 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/679a9ac2-d37f-4b13-bdf1-9705195350be-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467567 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467592 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/877a5c7e-5f1d-4404-b7ec-e8df15385dee-config\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467616 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqkh2\" (UniqueName: \"kubernetes.io/projected/6e31c209-d885-4235-87a7-972b045a6c72-kube-api-access-tqkh2\") pod \"control-plane-machine-set-operator-78cbb6b69f-ffc27\" (UID: \"6e31c209-d885-4235-87a7-972b045a6c72\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467640 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9621be61-f243-4e39-8b36-e7f8ccc3206d-config\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467663 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467692 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.467720 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/679a9ac2-d37f-4b13-bdf1-9705195350be-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468368 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38053e83-e26a-42e0-8a7c-7ebfd0714785-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468409 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-client\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468431 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4krb9\" (UniqueName: \"kubernetes.io/projected/682f968a-36d2-4da6-bc70-8cbb1eef8b8d-kube-api-access-4krb9\") pod \"ingress-canary-764ws\" (UID: \"682f968a-36d2-4da6-bc70-8cbb1eef8b8d\") " pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468452 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-metrics-certs\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468473 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c942617e-6e69-42e7-aa2b-62e634b14583-metrics-tls\") pod \"dns-operator-744455d44c-8jj25\" (UID: \"c942617e-6e69-42e7-aa2b-62e634b14583\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468497 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8ggt\" (UniqueName: \"kubernetes.io/projected/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-kube-api-access-q8ggt\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468532 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468558 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4743c41d-8037-4e64-bc07-befafeb23482-tmpfs\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468579 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-serving-cert\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468602 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38053e83-e26a-42e0-8a7c-7ebfd0714785-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468623 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/472551c7-c69c-4215-a7e8-22b9fc846cad-profile-collector-cert\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468647 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-trusted-ca\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468664 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-ca\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468685 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9621be61-f243-4e39-8b36-e7f8ccc3206d-serving-cert\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468708 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-csi-data-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468729 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-node-bootstrap-token\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.468749 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e37f501-ca1f-43d6-aace-f2a716cb5d79-metrics-tls\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.476973 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-certificates\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.477127 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-ca\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.478945 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-images\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.479381 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-client\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.479524 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38053e83-e26a-42e0-8a7c-7ebfd0714785-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.480229 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f5b6f36-f737-42b9-9403-db45098c682a-service-ca-bundle\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.482043 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-etcd-service-ca\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.483509 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-trusted-ca\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.483867 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-serving-cert\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.484345 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4743c41d-8037-4e64-bc07-befafeb23482-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.484660 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/679a9ac2-d37f-4b13-bdf1-9705195350be-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.484756 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.485166 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/679a9ac2-d37f-4b13-bdf1-9705195350be-config\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.485578 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/877a5c7e-5f1d-4404-b7ec-e8df15385dee-config\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.491091 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-config\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.491356 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4743c41d-8037-4e64-bc07-befafeb23482-webhook-cert\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.491939 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:33.991919681 +0000 UTC m=+147.322058561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.493336 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.493764 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/4743c41d-8037-4e64-bc07-befafeb23482-tmpfs\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.494116 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.495809 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.498293 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94b03a02-dc7e-43ba-8183-934fb6649cbf-srv-cert\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.500140 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.500184 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-proxy-tls\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.500381 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/472551c7-c69c-4215-a7e8-22b9fc846cad-srv-cert\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.500985 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/472551c7-c69c-4215-a7e8-22b9fc846cad-profile-collector-cert\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.503903 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/877a5c7e-5f1d-4404-b7ec-e8df15385dee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.515750 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-stats-auth\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522298 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38053e83-e26a-42e0-8a7c-7ebfd0714785-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522410 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522481 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-default-certificate\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522522 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2af6d859-3308-4f34-a2b8-cfb2f697bd46-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-b689q\" (UID: \"2af6d859-3308-4f34-a2b8-cfb2f697bd46\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522560 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-tls\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522593 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94b03a02-dc7e-43ba-8183-934fb6649cbf-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522872 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6e31c209-d885-4235-87a7-972b045a6c72-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ffc27\" (UID: \"6e31c209-d885-4235-87a7-972b045a6c72\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.522916 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-proxy-tls\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.523694 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c942617e-6e69-42e7-aa2b-62e634b14583-metrics-tls\") pod \"dns-operator-744455d44c-8jj25\" (UID: \"c942617e-6e69-42e7-aa2b-62e634b14583\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.524739 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m9xx\" (UniqueName: \"kubernetes.io/projected/c942617e-6e69-42e7-aa2b-62e634b14583-kube-api-access-8m9xx\") pod \"dns-operator-744455d44c-8jj25\" (UID: \"c942617e-6e69-42e7-aa2b-62e634b14583\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.527274 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f5b6f36-f737-42b9-9403-db45098c682a-metrics-certs\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.547806 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnwp4\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-kube-api-access-xnwp4\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.557357 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/877a5c7e-5f1d-4404-b7ec-e8df15385dee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mnjvx\" (UID: \"877a5c7e-5f1d-4404-b7ec-e8df15385dee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.560943 4679 generic.go:334] "Generic (PLEG): container finished" podID="3d1572f5-052f-4d3f-83d4-b631497c1d4f" containerID="eb8a791cd2f85f8701a782b1c4826e29e968303243455a88b1c802f4bd1aa1a3" exitCode=0 Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.561029 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" event={"ID":"3d1572f5-052f-4d3f-83d4-b631497c1d4f","Type":"ContainerDied","Data":"eb8a791cd2f85f8701a782b1c4826e29e968303243455a88b1c802f4bd1aa1a3"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.561075 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" event={"ID":"3d1572f5-052f-4d3f-83d4-b631497c1d4f","Type":"ContainerStarted","Data":"857eeebcacca766395b9a9011f5e28da6468e6d92e1c2e82f12a688d45b75e24"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.563055 4679 generic.go:334] "Generic (PLEG): container finished" podID="8fe15a15-98ee-4b83-98a6-1d09f577aae6" containerID="10ab21d9db87ecbeec3b412ad03eaed00c53c22ff94330cb68a6c2d1580a40a8" exitCode=0 Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.563272 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" event={"ID":"8fe15a15-98ee-4b83-98a6-1d09f577aae6","Type":"ContainerDied","Data":"10ab21d9db87ecbeec3b412ad03eaed00c53c22ff94330cb68a6c2d1580a40a8"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.563298 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" event={"ID":"8fe15a15-98ee-4b83-98a6-1d09f577aae6","Type":"ContainerStarted","Data":"662a2b08b4de008eb62ecf3de90cddd9d5a034969f32fdf3a414eec195a9e07a"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.564874 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" event={"ID":"e222153a-8cab-4ef0-8227-13be178f3f82","Type":"ContainerStarted","Data":"f3675a2ad712faee148fa3426c746848c0392a15b858d44458a61cb55e0ad237"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.564907 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" event={"ID":"e222153a-8cab-4ef0-8227-13be178f3f82","Type":"ContainerStarted","Data":"1920dac619ae3cc880cf0e6106f6f203f13728eedf86b806ff746882ac3ede1d"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.564920 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" event={"ID":"e222153a-8cab-4ef0-8227-13be178f3f82","Type":"ContainerStarted","Data":"eef5486311981bf46bb53840ff5d939e587d22cbcffb2a4ebdef18051abc4366"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.565530 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" event={"ID":"006c66c2-65fc-4cf7-9710-dfc64280b3c0","Type":"ContainerStarted","Data":"0b2f4a9e069566a27161a56e6bf3a7b6543e761bfc831e379257caee89ca0488"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.566099 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zr47g" event={"ID":"497dbe86-ba0e-40cb-a0ed-9608281cc6b2","Type":"ContainerStarted","Data":"10b99745d547ea9133902a724bb1a622c6483b902779dbca459376a039ef9246"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.567078 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" event={"ID":"9821b1bc-27fb-4720-afce-16072b779105","Type":"ContainerStarted","Data":"3381b86a8f7f228e39088ef40f6975e0504d2e4766c9f34cc54aeabbad8ce9d6"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.567111 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" event={"ID":"9821b1bc-27fb-4720-afce-16072b779105","Type":"ContainerStarted","Data":"356febb2b813d8eee1812ae7ab6b8b90d9336a3cf0dda7aaf0d49085ddd1f085"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.567370 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.568225 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" event={"ID":"7caf2887-b32b-4d3d-9c93-a116076dde2c","Type":"ContainerStarted","Data":"6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.568256 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" event={"ID":"7caf2887-b32b-4d3d-9c93-a116076dde2c","Type":"ContainerStarted","Data":"ba18f9ff04c3e78332adcbe93af0667c17e6f0b70d95597e341599be4072bce0"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.568774 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.569989 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.570132 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.070102825 +0000 UTC m=+147.400241685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570181 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chvjg\" (UniqueName: \"kubernetes.io/projected/9621be61-f243-4e39-8b36-e7f8ccc3206d-kube-api-access-chvjg\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570210 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d40d1aca-74a1-44e7-aed1-c595ce552803-config-volume\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570229 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e4fa17b-058b-46c2-8383-fc1fd5ae9b56-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9fzfq\" (UID: \"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570252 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e37f501-ca1f-43d6-aace-f2a716cb5d79-config-volume\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570274 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq5k8\" (UniqueName: \"kubernetes.io/projected/d40d1aca-74a1-44e7-aed1-c595ce552803-kube-api-access-kq5k8\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570349 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570376 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5jr\" (UniqueName: \"kubernetes.io/projected/d079b995-5285-4810-912b-0f41cde4a9d9-kube-api-access-vl5jr\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570398 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-socket-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570419 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hstng\" (UniqueName: \"kubernetes.io/projected/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-kube-api-access-hstng\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570440 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-signing-cabundle\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570455 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-plugins-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570479 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz5hf\" (UniqueName: \"kubernetes.io/projected/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-kube-api-access-wz5hf\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570497 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/9e4fa17b-058b-46c2-8383-fc1fd5ae9b56-kube-api-access-c4qfb\") pod \"package-server-manager-789f6589d5-9fzfq\" (UID: \"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570514 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-certs\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570538 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-registration-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570554 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-mountpoint-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570569 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhb88\" (UniqueName: \"kubernetes.io/projected/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-kube-api-access-nhb88\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570588 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lddl\" (UniqueName: \"kubernetes.io/projected/1e37f501-ca1f-43d6-aace-f2a716cb5d79-kube-api-access-2lddl\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570603 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-signing-key\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570671 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9621be61-f243-4e39-8b36-e7f8ccc3206d-config\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570686 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570704 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4krb9\" (UniqueName: \"kubernetes.io/projected/682f968a-36d2-4da6-bc70-8cbb1eef8b8d-kube-api-access-4krb9\") pod \"ingress-canary-764ws\" (UID: \"682f968a-36d2-4da6-bc70-8cbb1eef8b8d\") " pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570732 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.570850 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-plugins-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.571125 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9621be61-f243-4e39-8b36-e7f8ccc3206d-serving-cert\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.571154 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-csi-data-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.571170 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-node-bootstrap-token\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.571187 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e37f501-ca1f-43d6-aace-f2a716cb5d79-metrics-tls\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.571203 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d40d1aca-74a1-44e7-aed1-c595ce552803-secret-volume\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.571220 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/682f968a-36d2-4da6-bc70-8cbb1eef8b8d-cert\") pod \"ingress-canary-764ws\" (UID: \"682f968a-36d2-4da6-bc70-8cbb1eef8b8d\") " pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.571371 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.071358165 +0000 UTC m=+147.401497095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.572361 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e37f501-ca1f-43d6-aace-f2a716cb5d79-config-volume\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.572364 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9621be61-f243-4e39-8b36-e7f8ccc3206d-config\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.572595 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-csi-data-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.572698 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d40d1aca-74a1-44e7-aed1-c595ce552803-config-volume\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.573177 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-signing-cabundle\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.574430 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" event={"ID":"5c99d4c8-d743-4050-bf4b-1710297d1c4f","Type":"ContainerStarted","Data":"78ea15af7739ab9e0357763d94edeeb456451bf5d36537af8265e38e08753dd9"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.574467 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" event={"ID":"5c99d4c8-d743-4050-bf4b-1710297d1c4f","Type":"ContainerStarted","Data":"38b4a59349a8fa5f1c9096192075ea91cdea6f2dc88a2c2ad675af251b296702"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.574495 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.574671 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-socket-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.574720 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-mountpoint-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.574726 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d079b995-5285-4810-912b-0f41cde4a9d9-registration-dir\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.575732 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.575938 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-signing-key\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.576384 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38053e83-e26a-42e0-8a7c-7ebfd0714785-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-j2f7q\" (UID: \"38053e83-e26a-42e0-8a7c-7ebfd0714785\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.576571 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" event={"ID":"d91d02e1-27b2-41c4-aedd-8199c2f6eb42","Type":"ContainerStarted","Data":"e050e028916cfc6f5a9f374bf73975f80626eb1af48a28dfc1b3539e495756f7"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.576616 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" event={"ID":"d91d02e1-27b2-41c4-aedd-8199c2f6eb42","Type":"ContainerStarted","Data":"ee312993f7425eccb880fcfa888e6894dd23ca49017298035301ca8e3b0f7ac0"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.576621 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9621be61-f243-4e39-8b36-e7f8ccc3206d-serving-cert\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.577622 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" event={"ID":"143f3a78-2efe-4004-8bc6-fdab837914c1","Type":"ContainerStarted","Data":"8ed6354f5dc3bffa010f2f8865e7e3f2b90e0b21705e78ab33e8d01c80986c87"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.579361 4679 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-cx5cx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.579446 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" podUID="9821b1bc-27fb-4720-afce-16072b779105" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.579463 4679 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-p49dd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.579465 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bk2f" event={"ID":"9d97c73c-99b7-452c-9e31-f7c99d4c93f3","Type":"ContainerStarted","Data":"1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892"} Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.579488 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" podUID="7caf2887-b32b-4d3d-9c93-a116076dde2c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.580283 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1e37f501-ca1f-43d6-aace-f2a716cb5d79-metrics-tls\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.581589 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-certs\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.581679 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-node-bootstrap-token\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.585085 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e4fa17b-058b-46c2-8383-fc1fd5ae9b56-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9fzfq\" (UID: \"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.585258 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/682f968a-36d2-4da6-bc70-8cbb1eef8b8d-cert\") pod \"ingress-canary-764ws\" (UID: \"682f968a-36d2-4da6-bc70-8cbb1eef8b8d\") " pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.585494 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d40d1aca-74a1-44e7-aed1-c595ce552803-secret-volume\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.595226 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltrlw\" (UniqueName: \"kubernetes.io/projected/0be4a61e-cc1b-478d-b6dd-c602f29dab5d-kube-api-access-ltrlw\") pod \"machine-config-controller-84d6567774-lbr2v\" (UID: \"0be4a61e-cc1b-478d-b6dd-c602f29dab5d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.610029 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.628409 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhqh4\" (UniqueName: \"kubernetes.io/projected/0f5b6f36-f737-42b9-9403-db45098c682a-kube-api-access-jhqh4\") pod \"router-default-5444994796-bf564\" (UID: \"0f5b6f36-f737-42b9-9403-db45098c682a\") " pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.631367 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.643928 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88csv\" (UniqueName: \"kubernetes.io/projected/94b03a02-dc7e-43ba-8183-934fb6649cbf-kube-api-access-88csv\") pod \"olm-operator-6b444d44fb-mr8tq\" (UID: \"94b03a02-dc7e-43ba-8183-934fb6649cbf\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.652379 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.652674 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-964qh\" (UniqueName: \"kubernetes.io/projected/2af6d859-3308-4f34-a2b8-cfb2f697bd46-kube-api-access-964qh\") pod \"multus-admission-controller-857f4d67dd-b689q\" (UID: \"2af6d859-3308-4f34-a2b8-cfb2f697bd46\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.659927 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.671951 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.675832 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.175810952 +0000 UTC m=+147.505949822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.676355 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmmpk\" (UniqueName: \"kubernetes.io/projected/4743c41d-8037-4e64-bc07-befafeb23482-kube-api-access-vmmpk\") pod \"packageserver-d55dfcdfc-rmsxd\" (UID: \"4743c41d-8037-4e64-bc07-befafeb23482\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.688455 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.695614 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vnw7n"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.703173 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-bound-sa-token\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.705803 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zh5z\" (UniqueName: \"kubernetes.io/projected/ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06-kube-api-access-6zh5z\") pod \"kube-storage-version-migrator-operator-b67b599dd-5xxrk\" (UID: \"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.707978 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.723982 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.731996 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/679a9ac2-d37f-4b13-bdf1-9705195350be-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kmk7d\" (UID: \"679a9ac2-d37f-4b13-bdf1-9705195350be\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.739822 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.748194 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.750651 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqkh2\" (UniqueName: \"kubernetes.io/projected/6e31c209-d885-4235-87a7-972b045a6c72-kube-api-access-tqkh2\") pod \"control-plane-machine-set-operator-78cbb6b69f-ffc27\" (UID: \"6e31c209-d885-4235-87a7-972b045a6c72\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.774011 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnkzw\" (UniqueName: \"kubernetes.io/projected/472551c7-c69c-4215-a7e8-22b9fc846cad-kube-api-access-mnkzw\") pod \"catalog-operator-68c6474976-pp8hr\" (UID: \"472551c7-c69c-4215-a7e8-22b9fc846cad\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.777272 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.777315 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj"] Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.777649 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.277626087 +0000 UTC m=+147.607764947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.788075 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlw27\" (UniqueName: \"kubernetes.io/projected/95274f18-9c03-4d32-b14f-621646f64534-kube-api-access-jlw27\") pod \"migrator-59844c95c7-lzh87\" (UID: \"95274f18-9c03-4d32-b14f-621646f64534\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.822030 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8ggt\" (UniqueName: \"kubernetes.io/projected/3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5-kube-api-access-q8ggt\") pod \"machine-config-operator-74547568cd-dzxkm\" (UID: \"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.822162 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-68gpm"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.853216 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k96fw\" (UniqueName: \"kubernetes.io/projected/d29c8fe4-ed06-47c3-a2af-132f3cef0e50-kube-api-access-k96fw\") pod \"etcd-operator-b45778765-vmxn8\" (UID: \"d29c8fe4-ed06-47c3-a2af-132f3cef0e50\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.881936 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.882500 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.382479863 +0000 UTC m=+147.712618723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.882494 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chvjg\" (UniqueName: \"kubernetes.io/projected/9621be61-f243-4e39-8b36-e7f8ccc3206d-kube-api-access-chvjg\") pod \"service-ca-operator-777779d784-h55w5\" (UID: \"9621be61-f243-4e39-8b36-e7f8ccc3206d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.884434 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.893178 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq5k8\" (UniqueName: \"kubernetes.io/projected/d40d1aca-74a1-44e7-aed1-c595ce552803-kube-api-access-kq5k8\") pod \"collect-profiles-29411175-gkfrf\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:33 crc kubenswrapper[4679]: W1202 10:21:33.918808 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod975c5cdb_bf9b_4598_9b66_7b4d1e6ebc76.slice/crio-ed6a729ba2d992773cd58da28e07d48ccb1ab06a1e1db492907f27af8d2d794d WatchSource:0}: Error finding container ed6a729ba2d992773cd58da28e07d48ccb1ab06a1e1db492907f27af8d2d794d: Status 404 returned error can't find the container with id ed6a729ba2d992773cd58da28e07d48ccb1ab06a1e1db492907f27af8d2d794d Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.927259 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhb88\" (UniqueName: \"kubernetes.io/projected/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-kube-api-access-nhb88\") pod \"marketplace-operator-79b997595-4hmbk\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.940011 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.956520 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hstng\" (UniqueName: \"kubernetes.io/projected/5632ae04-eafc-4d03-bbbd-0cd0e8338c71-kube-api-access-hstng\") pod \"machine-config-server-zl899\" (UID: \"5632ae04-eafc-4d03-bbbd-0cd0e8338c71\") " pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.964159 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qcmqt"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.965769 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.966928 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lddl\" (UniqueName: \"kubernetes.io/projected/1e37f501-ca1f-43d6-aace-f2a716cb5d79-kube-api-access-2lddl\") pod \"dns-default-x9sld\" (UID: \"1e37f501-ca1f-43d6-aace-f2a716cb5d79\") " pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.977420 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.982803 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8jj25"] Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.983835 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:33 crc kubenswrapper[4679]: E1202 10:21:33.984483 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.484462701 +0000 UTC m=+147.814601561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:33 crc kubenswrapper[4679]: I1202 10:21:33.984875 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5jr\" (UniqueName: \"kubernetes.io/projected/d079b995-5285-4810-912b-0f41cde4a9d9-kube-api-access-vl5jr\") pod \"csi-hostpathplugin-bbxkc\" (UID: \"d079b995-5285-4810-912b-0f41cde4a9d9\") " pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.000525 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.008262 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz5hf\" (UniqueName: \"kubernetes.io/projected/3252b1c3-e4b8-4b96-9e73-dc889b04ccb3-kube-api-access-wz5hf\") pod \"service-ca-9c57cc56f-kvtft\" (UID: \"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.013666 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4krb9\" (UniqueName: \"kubernetes.io/projected/682f968a-36d2-4da6-bc70-8cbb1eef8b8d-kube-api-access-4krb9\") pod \"ingress-canary-764ws\" (UID: \"682f968a-36d2-4da6-bc70-8cbb1eef8b8d\") " pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.016933 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.030964 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.043661 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/9e4fa17b-058b-46c2-8383-fc1fd5ae9b56-kube-api-access-c4qfb\") pod \"package-server-manager-789f6589d5-9fzfq\" (UID: \"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:34 crc kubenswrapper[4679]: W1202 10:21:34.044564 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-bda4d428328af6c4620c36c5fa9341d7109f838d76d608fa6361b624c1d52c6b WatchSource:0}: Error finding container bda4d428328af6c4620c36c5fa9341d7109f838d76d608fa6361b624c1d52c6b: Status 404 returned error can't find the container with id bda4d428328af6c4620c36c5fa9341d7109f838d76d608fa6361b624c1d52c6b Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.057902 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.067622 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.077907 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-764ws" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.084023 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-zl899" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.084473 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.084772 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.58475694 +0000 UTC m=+147.914895800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.093715 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.098133 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.106988 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.154009 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.154046 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.187072 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.187383 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.687370463 +0000 UTC m=+148.017509323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.288168 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.288665 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.788647015 +0000 UTC m=+148.118785875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.298116 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk"] Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.322171 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx"] Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.390923 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.391230 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.891194635 +0000 UTC m=+148.221333495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.416153 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wppmh" podStartSLOduration=130.416128107 podStartE2EDuration="2m10.416128107s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:34.367471894 +0000 UTC m=+147.697610754" watchObservedRunningTime="2025-12-02 10:21:34.416128107 +0000 UTC m=+147.746266967" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.437457 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v"] Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.492993 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.493273 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:34.993258456 +0000 UTC m=+148.323397316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.585470 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5bk2f" podStartSLOduration=130.585446312 podStartE2EDuration="2m10.585446312s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:34.555084722 +0000 UTC m=+147.885223592" watchObservedRunningTime="2025-12-02 10:21:34.585446312 +0000 UTC m=+147.915585162" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.594990 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.595353 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.095339866 +0000 UTC m=+148.425478726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.666422 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq"] Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.672492 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" event={"ID":"006c66c2-65fc-4cf7-9710-dfc64280b3c0","Type":"ContainerStarted","Data":"e2957f2e2f2f5b2d690ba0da9f72f8f0bebcf73a2f2d772881f3d6f233da229d"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.692523 4679 generic.go:334] "Generic (PLEG): container finished" podID="143f3a78-2efe-4004-8bc6-fdab837914c1" containerID="d6875a5b3279c90d3612970f4653be628e6fb72764f9f485cd97d588a08f2d15" exitCode=0 Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.700053 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.700423 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.200403338 +0000 UTC m=+148.530542198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.700456 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" event={"ID":"143f3a78-2efe-4004-8bc6-fdab837914c1","Type":"ContainerDied","Data":"d6875a5b3279c90d3612970f4653be628e6fb72764f9f485cd97d588a08f2d15"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.718587 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" event={"ID":"877a5c7e-5f1d-4404-b7ec-e8df15385dee","Type":"ContainerStarted","Data":"afa0664900b80518aed7fb9ba7a71491e90c0c7fa63a0b4c7166d74bb10db1fb"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.731368 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-b689q"] Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.731571 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" event={"ID":"3142a842-274b-4193-ac44-19b98f0a3a94","Type":"ContainerStarted","Data":"2e4c7571cdfb6dfddfed0e8a594fa94bfb1fcb18aba25c81521a42310a9cb45a"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.758358 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" event={"ID":"8fe15a15-98ee-4b83-98a6-1d09f577aae6","Type":"ContainerStarted","Data":"9043ed963ddc84c98094ff1dec3b4264ac73e45816b04cf2de871f2b7ff6011f"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.758455 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.760486 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" event={"ID":"c082626c-da29-486a-988f-c4566ee26c91","Type":"ContainerStarted","Data":"9a9583d28f74c4d05a27370be4e7c43c2674b690770994bf498f74800c931fd1"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.770142 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" event={"ID":"2df7188a-a756-44c4-988e-81fd270db7aa","Type":"ContainerStarted","Data":"36c796635e8c41bc7337119894ce924c40421b5aa350320c705c6abdb0d815b6"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.776843 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q"] Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.786288 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" event={"ID":"d91d02e1-27b2-41c4-aedd-8199c2f6eb42","Type":"ContainerStarted","Data":"b7f0f4ffa742e2e3eb4a0f95ef380d54290abaf1933e9d1f696d1a6cdf45cfa0"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.801967 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.803161 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.303118594 +0000 UTC m=+148.633257514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.815986 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bda4d428328af6c4620c36c5fa9341d7109f838d76d608fa6361b624c1d52c6b"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.837101 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" podStartSLOduration=130.837075499 podStartE2EDuration="2m10.837075499s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:34.834325374 +0000 UTC m=+148.164464244" watchObservedRunningTime="2025-12-02 10:21:34.837075499 +0000 UTC m=+148.167214359" Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.870478 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9a8692d5b06b547cd28499e7106cb6622d58a7a62ab71bc4c7c314b59fae25d3"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.902894 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:34 crc kubenswrapper[4679]: E1202 10:21:34.905785 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.405761828 +0000 UTC m=+148.735900698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.906918 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" event={"ID":"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76","Type":"ContainerStarted","Data":"ed6a729ba2d992773cd58da28e07d48ccb1ab06a1e1db492907f27af8d2d794d"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.979719 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zr47g" event={"ID":"497dbe86-ba0e-40cb-a0ed-9608281cc6b2","Type":"ContainerStarted","Data":"57a8d6951ef6b61fbe6a0bcf409ed29984ff5876463cddd36e61733fd51230f1"} Dec 02 10:21:34 crc kubenswrapper[4679]: I1202 10:21:34.981347 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.005604 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" event={"ID":"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06","Type":"ContainerStarted","Data":"34e567a60d0a5e54d796d68a0476ac50536735dfd0a53c38cc62d7f9ec8fa04c"} Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.005782 4679 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr47g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.005827 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr47g" podUID="497dbe86-ba0e-40cb-a0ed-9608281cc6b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.006432 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.006823 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.506804624 +0000 UTC m=+148.836943494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.012830 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" event={"ID":"c942617e-6e69-42e7-aa2b-62e634b14583","Type":"ContainerStarted","Data":"4cfe66ae19f80b4909eea2d38ebf86bffeaecf8a1d27528d1723b6731a5aa588"} Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.041616 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bf564" event={"ID":"0f5b6f36-f737-42b9-9403-db45098c682a","Type":"ContainerStarted","Data":"af2addb256a855eacd386bcaf355ff53b516e93975fa3ef33a4ee4ee2dfeb7b3"} Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.065556 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd"] Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.080198 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr"] Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.105575 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vmxn8"] Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.110224 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"44d75f9bac5524a440d6c7675f829f5fec1dd638dd8dea6b5e89a481580f02ed"} Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.112493 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.114139 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.614117859 +0000 UTC m=+148.944256739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.132919 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" event={"ID":"5ca62df6-c0e6-4872-9302-918bd0c4b39a","Type":"ContainerStarted","Data":"73cacce9bf6fe921c95e51b55dcd5951c9c52c89de5ea95c43e11180cfe5e5b5"} Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.132987 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27"] Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.147125 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.147211 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.172094 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" podStartSLOduration=130.172077033 podStartE2EDuration="2m10.172077033s" podCreationTimestamp="2025-12-02 10:19:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.170838604 +0000 UTC m=+148.500977464" watchObservedRunningTime="2025-12-02 10:21:35.172077033 +0000 UTC m=+148.502215893" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.223565 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.233933 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.733898179 +0000 UTC m=+149.064037089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.256737 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lsz66" podStartSLOduration=131.2567128 podStartE2EDuration="2m11.2567128s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.222783725 +0000 UTC m=+148.552922585" watchObservedRunningTime="2025-12-02 10:21:35.2567128 +0000 UTC m=+148.586851660" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.325609 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.825575603 +0000 UTC m=+149.155714483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.325866 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.326405 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.326767 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.826754661 +0000 UTC m=+149.156893531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.432899 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.433226 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:35.933206275 +0000 UTC m=+149.263345135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.536157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.537008 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.036991976 +0000 UTC m=+149.367130836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.637869 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.638231 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.138210747 +0000 UTC m=+149.468349607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.679063 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm"] Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.735348 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" podStartSLOduration=131.73533208 podStartE2EDuration="2m11.73533208s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.708962164 +0000 UTC m=+149.039101024" watchObservedRunningTime="2025-12-02 10:21:35.73533208 +0000 UTC m=+149.065470940" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.736648 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-zr47g" podStartSLOduration=131.736639601 podStartE2EDuration="2m11.736639601s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.73450098 +0000 UTC m=+149.064639840" watchObservedRunningTime="2025-12-02 10:21:35.736639601 +0000 UTC m=+149.066778471" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.739114 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.739419 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.239408176 +0000 UTC m=+149.569547036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.764374 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-nrwbn" podStartSLOduration=131.764353878 podStartE2EDuration="2m11.764353878s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.757468655 +0000 UTC m=+149.087607515" watchObservedRunningTime="2025-12-02 10:21:35.764353878 +0000 UTC m=+149.094492758" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.795724 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-z95rb" podStartSLOduration=131.795702091 podStartE2EDuration="2m11.795702091s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.793511719 +0000 UTC m=+149.123650589" watchObservedRunningTime="2025-12-02 10:21:35.795702091 +0000 UTC m=+149.125840951" Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.816139 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq"] Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.839885 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.839991 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.339969421 +0000 UTC m=+149.670108291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.840283 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.840752 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.340741719 +0000 UTC m=+149.670880579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.927205 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7psth" podStartSLOduration=131.927184649 podStartE2EDuration="2m11.927184649s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:35.925130501 +0000 UTC m=+149.255269361" watchObservedRunningTime="2025-12-02 10:21:35.927184649 +0000 UTC m=+149.257323509" Dec 02 10:21:35 crc kubenswrapper[4679]: W1202 10:21:35.931726 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c36e0f5_8ec5_42ed_a9a3_71ae29e5ddd5.slice/crio-c9b8a656a19b3e09b480a5aa214c95bdedbc1e9b639fc3e721cf5e441c88084b WatchSource:0}: Error finding container c9b8a656a19b3e09b480a5aa214c95bdedbc1e9b639fc3e721cf5e441c88084b: Status 404 returned error can't find the container with id c9b8a656a19b3e09b480a5aa214c95bdedbc1e9b639fc3e721cf5e441c88084b Dec 02 10:21:35 crc kubenswrapper[4679]: I1202 10:21:35.940979 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:35 crc kubenswrapper[4679]: E1202 10:21:35.941346 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.441327425 +0000 UTC m=+149.771466285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.043641 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.044242 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.544227695 +0000 UTC m=+149.874366555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.099275 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.143132 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x9sld"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.147005 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.147528 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.647280668 +0000 UTC m=+149.977419548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.210602 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.260982 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.262004 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.761990219 +0000 UTC m=+150.092129079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.262278 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" event={"ID":"472551c7-c69c-4215-a7e8-22b9fc846cad","Type":"ContainerStarted","Data":"187158a827f7b09e3c36ef0dce28f5394ead9602cd752c03669a619ddce76902"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.286058 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" event={"ID":"0be4a61e-cc1b-478d-b6dd-c602f29dab5d","Type":"ContainerStarted","Data":"dec15514e3e600d735372341d64e4ce695fd8f06a9b94d76ff65e0973e9f9c8b"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.286121 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" event={"ID":"0be4a61e-cc1b-478d-b6dd-c602f29dab5d","Type":"ContainerStarted","Data":"48b8bc2ea362ca79da0c1960d0fb7b231c4a0af0801d3483ab9ec1e9f62a2a15"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.345132 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7ba77ff06cc949f2bea8064381600ce5cabfcd78bc203e9a6f79f3fa5ab0266d"} Dec 02 10:21:36 crc kubenswrapper[4679]: W1202 10:21:36.345199 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd40d1aca_74a1_44e7_aed1_c595ce552803.slice/crio-dd12930a522795e7837343b4f05466a7a2080a23ff7e96792e9107bc44e51748 WatchSource:0}: Error finding container dd12930a522795e7837343b4f05466a7a2080a23ff7e96792e9107bc44e51748: Status 404 returned error can't find the container with id dd12930a522795e7837343b4f05466a7a2080a23ff7e96792e9107bc44e51748 Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.364179 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.364465 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.864450508 +0000 UTC m=+150.194589368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.390005 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kvtft"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.401525 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hmbk"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.403979 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.417480 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6c8fc2b36ed3b8d227a5d195db14ffe290696572894ad537bd8cff1c5cd0b909"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.418525 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:21:36 crc kubenswrapper[4679]: W1202 10:21:36.433196 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3252b1c3_e4b8_4b96_9e73_dc889b04ccb3.slice/crio-2725cb09174d3ab69afa4c69100987bb151a5d3452f97935ee896e3b8f4f671a WatchSource:0}: Error finding container 2725cb09174d3ab69afa4c69100987bb151a5d3452f97935ee896e3b8f4f671a: Status 404 returned error can't find the container with id 2725cb09174d3ab69afa4c69100987bb151a5d3452f97935ee896e3b8f4f671a Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.453104 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" event={"ID":"38053e83-e26a-42e0-8a7c-7ebfd0714785","Type":"ContainerStarted","Data":"d3548a347c98ed8a2fa228426de400923f04a9ea14eaff797bef3bc71ef48ee1"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.465372 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.465687 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:36.965668659 +0000 UTC m=+150.295807519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.467136 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h55w5"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.496288 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" event={"ID":"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56","Type":"ContainerStarted","Data":"d78a9d420e436871292e1ea4a84373eadd52488a677b002becd70adc3d869ffe"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.519551 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" event={"ID":"6e31c209-d885-4235-87a7-972b045a6c72","Type":"ContainerStarted","Data":"dd97c02b0a19b9efa6a6b920b62110f00b964568b386ad651756dd02e697ef40"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.533847 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bbxkc"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.541643 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zl899" event={"ID":"5632ae04-eafc-4d03-bbbd-0cd0e8338c71","Type":"ContainerStarted","Data":"10a37e4ff0932e419973ae67457e5322b93aad6afab10aeb015b6c893c0c837f"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.553240 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" event={"ID":"94b03a02-dc7e-43ba-8183-934fb6649cbf","Type":"ContainerStarted","Data":"2c101ad014ac2dd7fc65f6e3255815d172d909e736d8ad1db18fba8a55f96110"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.561432 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" event={"ID":"2df7188a-a756-44c4-988e-81fd270db7aa","Type":"ContainerStarted","Data":"819cbd91946b44984464fb085dfb24163c26f3baf4a80eabd34cd2432fc11d7e"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.563711 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.564414 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" event={"ID":"d29c8fe4-ed06-47c3-a2af-132f3cef0e50","Type":"ContainerStarted","Data":"78fbbd9c48edda3188468dfa723dcb1df9b39c6f68a8a8ddb687fe49cf4c9412"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.566097 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.566418 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.066403717 +0000 UTC m=+150.396542577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.572224 4679 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-68gpm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" start-of-body= Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.572274 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.18:6443/healthz\": dial tcp 10.217.0.18:6443: connect: connection refused" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.586449 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" event={"ID":"5ca62df6-c0e6-4872-9302-918bd0c4b39a","Type":"ContainerStarted","Data":"2719d35585fb8674fcdd2cf3792268e44e8905343072fa113c59bc5c9244693f"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.603604 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" podStartSLOduration=132.603591189 podStartE2EDuration="2m12.603591189s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:36.598726464 +0000 UTC m=+149.928865324" watchObservedRunningTime="2025-12-02 10:21:36.603591189 +0000 UTC m=+149.933730049" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.613155 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" event={"ID":"4743c41d-8037-4e64-bc07-befafeb23482","Type":"ContainerStarted","Data":"6a3d3cd3aeefcb9cd5729c407479a783960184b9341225deca88d4d64db5cff8"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.615289 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" event={"ID":"2af6d859-3308-4f34-a2b8-cfb2f697bd46","Type":"ContainerStarted","Data":"7de13aee7afae845b1ab51907f045be9c507874fb30ed9ddff38b711dc90bce9"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.626956 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" event={"ID":"3d1572f5-052f-4d3f-83d4-b631497c1d4f","Type":"ContainerStarted","Data":"e304191f340372beda5ad3a40c439d21a52a6ebb33f3ce11a9f194736b571a95"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.636993 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" event={"ID":"3142a842-274b-4193-ac44-19b98f0a3a94","Type":"ContainerStarted","Data":"8a683d59dfb9792f2adcca3674ba9b039b2be52951422b8999f115a7fce18995"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.638628 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.639238 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-764ws"] Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.646782 4679 patch_prober.go:28] interesting pod/console-operator-58897d9998-vnw7n container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.646829 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" podUID="3142a842-274b-4193-ac44-19b98f0a3a94" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.652605 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d35fbb6631fc3ad05097c5f0cc5f9c3a528d068b2cc9c6eadcda9c24392ac87c"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.659990 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bf564" event={"ID":"0f5b6f36-f737-42b9-9403-db45098c682a","Type":"ContainerStarted","Data":"7701c6204d48f85bcf2866f0266e06fc5d8fd0c4ca851953196e517de0d83516"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.669346 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.671849 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.171819607 +0000 UTC m=+150.501958547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.672752 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" event={"ID":"c082626c-da29-486a-988f-c4566ee26c91","Type":"ContainerStarted","Data":"d03cb2f70f3131e74bb67bc3564019d0e3fd7f027c23072267bfb93cdc15e612"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.673339 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qcmqt" podStartSLOduration=132.673327293 podStartE2EDuration="2m12.673327293s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:36.668670652 +0000 UTC m=+149.998809522" watchObservedRunningTime="2025-12-02 10:21:36.673327293 +0000 UTC m=+150.003466153" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.677032 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" event={"ID":"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5","Type":"ContainerStarted","Data":"c9b8a656a19b3e09b480a5aa214c95bdedbc1e9b639fc3e721cf5e441c88084b"} Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.678821 4679 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr47g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.678876 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr47g" podUID="497dbe86-ba0e-40cb-a0ed-9608281cc6b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.733363 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bf564" podStartSLOduration=132.733344136 podStartE2EDuration="2m12.733344136s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:36.731114333 +0000 UTC m=+150.061253193" watchObservedRunningTime="2025-12-02 10:21:36.733344136 +0000 UTC m=+150.063482996" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.770863 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.771005 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.270981189 +0000 UTC m=+150.601120049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.771166 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.773096 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.273077238 +0000 UTC m=+150.603216188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.798495 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" podStartSLOduration=132.798476741 podStartE2EDuration="2m12.798476741s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:36.754227321 +0000 UTC m=+150.084366191" watchObservedRunningTime="2025-12-02 10:21:36.798476741 +0000 UTC m=+150.128615601" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.829456 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" podStartSLOduration=132.829439735 podStartE2EDuration="2m12.829439735s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:36.800433617 +0000 UTC m=+150.130572477" watchObservedRunningTime="2025-12-02 10:21:36.829439735 +0000 UTC m=+150.159578595" Dec 02 10:21:36 crc kubenswrapper[4679]: I1202 10:21:36.898530 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:36 crc kubenswrapper[4679]: E1202 10:21:36.898961 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.398941033 +0000 UTC m=+150.729079893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.000600 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.001048 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.501032644 +0000 UTC m=+150.831171504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.102416 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.102774 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.602760476 +0000 UTC m=+150.932899336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.205090 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.205458 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.705444021 +0000 UTC m=+151.035582881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.305904 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.306609 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.8065884 +0000 UTC m=+151.136727260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.407245 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.407624 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:37.907612525 +0000 UTC m=+151.237751385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.509180 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.509445 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.0094184 +0000 UTC m=+151.339557250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.509536 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.509968 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.009951472 +0000 UTC m=+151.340090522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.611073 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.611448 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.111430659 +0000 UTC m=+151.441569519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.633721 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.641489 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:37 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:37 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:37 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.641571 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.696979 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" event={"ID":"3d1572f5-052f-4d3f-83d4-b631497c1d4f","Type":"ContainerStarted","Data":"3827073c87616b97daaf209f5c9a60a3757dd195c74a5caff7eed7475bea4c1b"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.703212 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" event={"ID":"94b03a02-dc7e-43ba-8183-934fb6649cbf","Type":"ContainerStarted","Data":"82ed1dec13d0ff9cf140e98b60bf6a09727bdb7ce06611a7f425847f0bcdcb77"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.703766 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.710220 4679 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-mr8tq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.710268 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" podUID="94b03a02-dc7e-43ba-8183-934fb6649cbf" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.710659 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" event={"ID":"2df7188a-a756-44c4-988e-81fd270db7aa","Type":"ContainerStarted","Data":"834a55adbad3b71e85b2c87f5f27199a794560830bfb6456957935c804edd5cd"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.713065 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.713458 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.213441728 +0000 UTC m=+151.543580588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.723240 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" event={"ID":"38053e83-e26a-42e0-8a7c-7ebfd0714785","Type":"ContainerStarted","Data":"2eab09784a12b502ec4094a4acb9a0205e76f699f94fcf86e91fff04a8abb32f"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.724073 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" podStartSLOduration=133.724054189 podStartE2EDuration="2m13.724054189s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:37.723479466 +0000 UTC m=+151.053618326" watchObservedRunningTime="2025-12-02 10:21:37.724054189 +0000 UTC m=+151.054193049" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.733472 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" event={"ID":"c942617e-6e69-42e7-aa2b-62e634b14583","Type":"ContainerStarted","Data":"4361aef95cc35e84a0464a605c7cd7b561e2b4b75b088efcabe969c86a4d4ae4"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.748790 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-j2f7q" podStartSLOduration=133.748771786 podStartE2EDuration="2m13.748771786s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:37.746810869 +0000 UTC m=+151.076949729" watchObservedRunningTime="2025-12-02 10:21:37.748771786 +0000 UTC m=+151.078910646" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.753612 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" event={"ID":"6e31c209-d885-4235-87a7-972b045a6c72","Type":"ContainerStarted","Data":"4bc84af010ca67223a0f0ab1d586e1481acd22e0e2a8270f64be862d1cd41266"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.766733 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" event={"ID":"a180dc8a-e1e6-4da9-bf62-e973ef9d4611","Type":"ContainerStarted","Data":"f09cea1eb410b7707a4fe11c8b8df1052a5fd0bfabcc95d71f61365de9beae3a"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.766780 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" event={"ID":"a180dc8a-e1e6-4da9-bf62-e973ef9d4611","Type":"ContainerStarted","Data":"8442f2ad5edda06d137c820f12bbea689c4fd8bfddb44ff703b570b12fcfa808"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.776268 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x9sld" event={"ID":"1e37f501-ca1f-43d6-aace-f2a716cb5d79","Type":"ContainerStarted","Data":"353fdcca97514b80dc31efdd521d395b4d2b0b2d4afe22ed349bdac117e8e98a"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.782467 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x9sld" event={"ID":"1e37f501-ca1f-43d6-aace-f2a716cb5d79","Type":"ContainerStarted","Data":"a9dc024d246c0092ad66c612205e7824e6c978929fec18b31a6aead460f0b93c"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.814410 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.816628 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.316605634 +0000 UTC m=+151.646744504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.818008 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" event={"ID":"143f3a78-2efe-4004-8bc6-fdab837914c1","Type":"ContainerStarted","Data":"ddf3f193fdea4f5744f502156493f29bc1d855f31af807fcaf6e6c6a4bae2c0a"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.842288 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" event={"ID":"d079b995-5285-4810-912b-0f41cde4a9d9","Type":"ContainerStarted","Data":"974dd9f7be7e92dc76b9741becae59ef6e93210335dec1603d77b500742186b3"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.856392 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" event={"ID":"679a9ac2-d37f-4b13-bdf1-9705195350be","Type":"ContainerStarted","Data":"67909e829ee5356d2ab7a8425cc637b3d2236c1cdf13138d67db487912c42f9e"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.857452 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" podStartSLOduration=133.857429432 podStartE2EDuration="2m13.857429432s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:37.856124931 +0000 UTC m=+151.186263831" watchObservedRunningTime="2025-12-02 10:21:37.857429432 +0000 UTC m=+151.187568302" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.861147 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ffc27" podStartSLOduration=133.86113296 podStartE2EDuration="2m13.86113296s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:37.787098844 +0000 UTC m=+151.117237704" watchObservedRunningTime="2025-12-02 10:21:37.86113296 +0000 UTC m=+151.191271830" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.879648 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" event={"ID":"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3","Type":"ContainerStarted","Data":"fd79074e28a670b2c20ef80407667c3b056d7a3049447330d98b90ffac885e35"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.879750 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" event={"ID":"3252b1c3-e4b8-4b96-9e73-dc889b04ccb3","Type":"ContainerStarted","Data":"2725cb09174d3ab69afa4c69100987bb151a5d3452f97935ee896e3b8f4f671a"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.894730 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" event={"ID":"d40d1aca-74a1-44e7-aed1-c595ce552803","Type":"ContainerStarted","Data":"63bc391a21663da67a2d4db4ae4cf6f7d60cd5513067324208f7e9e29f0a43c8"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.894789 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" event={"ID":"d40d1aca-74a1-44e7-aed1-c595ce552803","Type":"ContainerStarted","Data":"dd12930a522795e7837343b4f05466a7a2080a23ff7e96792e9107bc44e51748"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.906336 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-zl899" event={"ID":"5632ae04-eafc-4d03-bbbd-0cd0e8338c71","Type":"ContainerStarted","Data":"ae69b2594ec9208491872846d0ced8a86ccaeea47da58f0c2363c801d9b43498"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.909130 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" event={"ID":"2af6d859-3308-4f34-a2b8-cfb2f697bd46","Type":"ContainerStarted","Data":"61c1be0f01a5314dd67d5f8448764863f6b936155109c9087a9285a98ecd3f15"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.918236 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:37 crc kubenswrapper[4679]: E1202 10:21:37.918521 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.418509571 +0000 UTC m=+151.748648431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.926209 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" event={"ID":"95274f18-9c03-4d32-b14f-621646f64534","Type":"ContainerStarted","Data":"f34b3daf36ba855db2922ad19056c6a0603105ae21efc794e1b79c50ed3f1ed8"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.926260 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" event={"ID":"95274f18-9c03-4d32-b14f-621646f64534","Type":"ContainerStarted","Data":"fe6a289dedafa61c269c140f8765d93d66d2dc40514d1c4f77db917dcb081967"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.932359 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" event={"ID":"9621be61-f243-4e39-8b36-e7f8ccc3206d","Type":"ContainerStarted","Data":"73ab921f5d44e8f3e61d17ab89f3fa4abf742a95d0ade18e04a9fd776e0a261f"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.932416 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" event={"ID":"9621be61-f243-4e39-8b36-e7f8ccc3206d","Type":"ContainerStarted","Data":"52fd42a0292fef2c0a73b28e9e6f421e055736879c2c135275a785aac8922e4a"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.932807 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-zl899" podStartSLOduration=7.932792839 podStartE2EDuration="7.932792839s" podCreationTimestamp="2025-12-02 10:21:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:37.930879014 +0000 UTC m=+151.261017874" watchObservedRunningTime="2025-12-02 10:21:37.932792839 +0000 UTC m=+151.262931699" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.942425 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" event={"ID":"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56","Type":"ContainerStarted","Data":"b3b5e007e68e28b6ec8c7209e000583b2d3828eda86004d5d41de8f4bb9ee28b"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.946799 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mnjvx" event={"ID":"877a5c7e-5f1d-4404-b7ec-e8df15385dee","Type":"ContainerStarted","Data":"8a6315c8fc06b8b13fe8d795bbc8c68002af1d8e67d48bbe6a04601e06a51343"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.955592 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" event={"ID":"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5","Type":"ContainerStarted","Data":"4ec7de82abe451e1c165d3e04c434a3884a1ab468cc3f81c7d5852e88f15cb63"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.960112 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" event={"ID":"ee06ce00-b2ec-47e1-b7ec-d94b4fac7b06","Type":"ContainerStarted","Data":"9624e543a38ff5b145f8866d48ae16097355f0c1fab4ff1ffd7f3d9356fe0bf2"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.972326 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-764ws" event={"ID":"682f968a-36d2-4da6-bc70-8cbb1eef8b8d","Type":"ContainerStarted","Data":"94ecbeb67399dfb681d85158766012280238e32d19ec63427ad8b13bb59997a4"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.974830 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" event={"ID":"c082626c-da29-486a-988f-c4566ee26c91","Type":"ContainerStarted","Data":"3c357b9f10d2d82fc726070074bd77860d3c38718beabc8eb973a26476849e30"} Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.989632 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5xxrk" podStartSLOduration=133.989611706 podStartE2EDuration="2m13.989611706s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:37.986868801 +0000 UTC m=+151.317007661" watchObservedRunningTime="2025-12-02 10:21:37.989611706 +0000 UTC m=+151.319750566" Dec 02 10:21:37 crc kubenswrapper[4679]: I1202 10:21:37.989991 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" event={"ID":"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76","Type":"ContainerStarted","Data":"5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e"} Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.010840 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.011914 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" event={"ID":"472551c7-c69c-4215-a7e8-22b9fc846cad","Type":"ContainerStarted","Data":"163685d2f6c139a5c05392681f72a72fc064f3396db0b8d12706f2d1d125a7a0"} Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.013073 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.020865 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.022708 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.52267814 +0000 UTC m=+151.852817040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.035495 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6lmxj" podStartSLOduration=134.035452423 podStartE2EDuration="2m14.035452423s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:38.033331812 +0000 UTC m=+151.363470672" watchObservedRunningTime="2025-12-02 10:21:38.035452423 +0000 UTC m=+151.365591293" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.050705 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" event={"ID":"4743c41d-8037-4e64-bc07-befafeb23482","Type":"ContainerStarted","Data":"7062cf143e88b91b956e60834238f50b92595e1c63b900b3b47bb4a6574164ac"} Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.051295 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.087370 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.087448 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.089133 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.106677 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-pp8hr" podStartSLOduration=134.106659491 podStartE2EDuration="2m14.106659491s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:38.10152765 +0000 UTC m=+151.431666510" watchObservedRunningTime="2025-12-02 10:21:38.106659491 +0000 UTC m=+151.436798351" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.115683 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" event={"ID":"d29c8fe4-ed06-47c3-a2af-132f3cef0e50","Type":"ContainerStarted","Data":"dba4ebc91e673123102ce01911212a2b326a366dc6737d21b2202124bf0861a3"} Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.125515 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.127071 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.627058845 +0000 UTC m=+151.957197705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.128855 4679 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr47g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.129016 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr47g" podUID="497dbe86-ba0e-40cb-a0ed-9608281cc6b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.227692 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.227913 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.727880046 +0000 UTC m=+152.058018926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.228540 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.229250 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.729217988 +0000 UTC m=+152.059356858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.261048 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" podStartSLOduration=134.261027092 podStartE2EDuration="2m14.261027092s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:38.200770493 +0000 UTC m=+151.530909373" watchObservedRunningTime="2025-12-02 10:21:38.261027092 +0000 UTC m=+151.591165952" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.297488 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vmxn8" podStartSLOduration=134.297468356 podStartE2EDuration="2m14.297468356s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:38.26473003 +0000 UTC m=+151.594868900" watchObservedRunningTime="2025-12-02 10:21:38.297468356 +0000 UTC m=+151.627607216" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.329653 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.329931 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.829917886 +0000 UTC m=+152.160056746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.435659 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.436394 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:38.93637956 +0000 UTC m=+152.266518420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.457319 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-vnw7n" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.536866 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.537213 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.037194131 +0000 UTC m=+152.367332991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.571660 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wnwnx" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.638952 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.639470 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.139458996 +0000 UTC m=+152.469597856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.652262 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:38 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:38 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:38 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.652356 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.740756 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.741009 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.240980623 +0000 UTC m=+152.571119483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.741324 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.741654 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.241645529 +0000 UTC m=+152.571784389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.842123 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.842314 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.342274065 +0000 UTC m=+152.672412925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.842721 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.843127 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.343106615 +0000 UTC m=+152.673245515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.944473 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.944714 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.444663113 +0000 UTC m=+152.774801983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:38 crc kubenswrapper[4679]: I1202 10:21:38.944803 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:38 crc kubenswrapper[4679]: E1202 10:21:38.945121 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.445107544 +0000 UTC m=+152.775246404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.058399 4679 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rmsxd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.058470 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" podUID="4743c41d-8037-4e64-bc07-befafeb23482" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.058984 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.059213 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.559176999 +0000 UTC m=+152.889315859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.059251 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.059578 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.559564318 +0000 UTC m=+152.889703178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.159746 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" event={"ID":"c942617e-6e69-42e7-aa2b-62e634b14583","Type":"ContainerStarted","Data":"44f9da21362121475d74680b969a8a7e61ec7762a2c8699df0c2f07f0b84a2f7"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.160847 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.161127 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.661100336 +0000 UTC m=+152.991239186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.161538 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.161936 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.661923495 +0000 UTC m=+152.992062365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.186651 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" event={"ID":"3c36e0f5-8ec5-42ed-a9a3-71ae29e5ddd5","Type":"ContainerStarted","Data":"bc6b4535bb57f4a2eeb8206eb5dad618a8fea839b026d9aaf65cd298d473cec7"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.190918 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" event={"ID":"0be4a61e-cc1b-478d-b6dd-c602f29dab5d","Type":"ContainerStarted","Data":"0a3734ed8816aea0679b2bd142b7d6ebb07c465ba263b74bb512c35aace8f9c7"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.201128 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" event={"ID":"2af6d859-3308-4f34-a2b8-cfb2f697bd46","Type":"ContainerStarted","Data":"209183f87f4f36abfecaf9ec05f181bcd4d426315e9b5aa72163d3f26e3f2f1f"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.213482 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" event={"ID":"d079b995-5285-4810-912b-0f41cde4a9d9","Type":"ContainerStarted","Data":"bf327cd57e780805b9ac51b1864a9fb3e7eb15cda00a8a094b01793a3285e6a1"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.221764 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dzxkm" podStartSLOduration=135.221739794 podStartE2EDuration="2m15.221739794s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.216203413 +0000 UTC m=+152.546342293" watchObservedRunningTime="2025-12-02 10:21:39.221739794 +0000 UTC m=+152.551878664" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.224064 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" event={"ID":"9e4fa17b-058b-46c2-8383-fc1fd5ae9b56","Type":"ContainerStarted","Data":"817a30210e433bd3ad6a5b864e863e8bca4a5c66a70789cbcbfca63031ba3516"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.225371 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x9sld" event={"ID":"1e37f501-ca1f-43d6-aace-f2a716cb5d79","Type":"ContainerStarted","Data":"cceecb57fe73ef926105f6d43c7a473b91fcac8f7b0ee97fa1109e891ac21706"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.234463 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" event={"ID":"95274f18-9c03-4d32-b14f-621646f64534","Type":"ContainerStarted","Data":"56a915387da4406626310d70a47492613468ae3fe560b0937388ba8b6ef4d535"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.245918 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lbr2v" podStartSLOduration=135.245898317 podStartE2EDuration="2m15.245898317s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.244044493 +0000 UTC m=+152.574183353" watchObservedRunningTime="2025-12-02 10:21:39.245898317 +0000 UTC m=+152.576037177" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.257535 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-764ws" event={"ID":"682f968a-36d2-4da6-bc70-8cbb1eef8b8d","Type":"ContainerStarted","Data":"f87ae6788f11f3fd38183430401e77b133d14ecc17deb4eff0544f57b6fdbc76"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.265088 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.265423 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.765401369 +0000 UTC m=+153.095540229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.270423 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" event={"ID":"679a9ac2-d37f-4b13-bdf1-9705195350be","Type":"ContainerStarted","Data":"749df8c68f283179b9e0f110b9e008ce290f825e9e733786a27359183ed7c1d0"} Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.286517 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lzh87" podStartSLOduration=135.286493039 podStartE2EDuration="2m15.286493039s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.276145614 +0000 UTC m=+152.606284474" watchObservedRunningTime="2025-12-02 10:21:39.286493039 +0000 UTC m=+152.616631899" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.288450 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mr8tq" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.306009 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" podStartSLOduration=135.305985042 podStartE2EDuration="2m15.305985042s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.305779737 +0000 UTC m=+152.635918597" watchObservedRunningTime="2025-12-02 10:21:39.305985042 +0000 UTC m=+152.636123902" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.326510 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h55w5" podStartSLOduration=134.326489338 podStartE2EDuration="2m14.326489338s" podCreationTimestamp="2025-12-02 10:19:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.322009602 +0000 UTC m=+152.652148472" watchObservedRunningTime="2025-12-02 10:21:39.326489338 +0000 UTC m=+152.656628198" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.368822 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.369107 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.869095908 +0000 UTC m=+153.199234768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.386522 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-764ws" podStartSLOduration=9.386506311 podStartE2EDuration="9.386506311s" podCreationTimestamp="2025-12-02 10:21:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.353086599 +0000 UTC m=+152.683225479" watchObservedRunningTime="2025-12-02 10:21:39.386506311 +0000 UTC m=+152.716645171" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.387184 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" podStartSLOduration=135.387180097 podStartE2EDuration="2m15.387180097s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.385523518 +0000 UTC m=+152.715662378" watchObservedRunningTime="2025-12-02 10:21:39.387180097 +0000 UTC m=+152.717318957" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.407638 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-kvtft" podStartSLOduration=134.407622382 podStartE2EDuration="2m14.407622382s" podCreationTimestamp="2025-12-02 10:19:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.405730107 +0000 UTC m=+152.735868967" watchObservedRunningTime="2025-12-02 10:21:39.407622382 +0000 UTC m=+152.737761232" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.435287 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" podStartSLOduration=135.435266107 podStartE2EDuration="2m15.435266107s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.434029388 +0000 UTC m=+152.764168258" watchObservedRunningTime="2025-12-02 10:21:39.435266107 +0000 UTC m=+152.765404967" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.469874 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.470332 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:39.970294548 +0000 UTC m=+153.300433408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.474050 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kmk7d" podStartSLOduration=135.474030417 podStartE2EDuration="2m15.474030417s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.449441803 +0000 UTC m=+152.779580953" watchObservedRunningTime="2025-12-02 10:21:39.474030417 +0000 UTC m=+152.804169277" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.483151 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xjmb2" podStartSLOduration=135.483134042 podStartE2EDuration="2m15.483134042s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:39.475660505 +0000 UTC m=+152.805799385" watchObservedRunningTime="2025-12-02 10:21:39.483134042 +0000 UTC m=+152.813272902" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.485205 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vr2r9"] Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.486242 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.488371 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.502563 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vr2r9"] Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.571468 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-utilities\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.571539 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-catalog-content\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.571598 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99btx\" (UniqueName: \"kubernetes.io/projected/91618844-4970-41c8-ba39-ed76f10a7ff5-kube-api-access-99btx\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.571640 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.571984 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.071969509 +0000 UTC m=+153.402108369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.636400 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:39 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:39 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:39 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.636471 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.672989 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.673341 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-utilities\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.673392 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-catalog-content\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.673449 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99btx\" (UniqueName: \"kubernetes.io/projected/91618844-4970-41c8-ba39-ed76f10a7ff5-kube-api-access-99btx\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.673667 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.17363726 +0000 UTC m=+153.503776130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.674262 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-utilities\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.674397 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-catalog-content\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.685831 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hjp6m"] Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.686753 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.689580 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.703816 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hjp6m"] Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.728360 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99btx\" (UniqueName: \"kubernetes.io/projected/91618844-4970-41c8-ba39-ed76f10a7ff5-kube-api-access-99btx\") pod \"community-operators-vr2r9\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.738104 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmsxd" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.774813 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqgzw\" (UniqueName: \"kubernetes.io/projected/1a86b0c7-517b-465f-bbe5-d957b687a38d-kube-api-access-jqgzw\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.774867 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-catalog-content\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.774917 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.774947 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-utilities\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.775285 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.27526031 +0000 UTC m=+153.605399170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.806942 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.866583 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.876995 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.877162 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqgzw\" (UniqueName: \"kubernetes.io/projected/1a86b0c7-517b-465f-bbe5-d957b687a38d-kube-api-access-jqgzw\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.877197 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-catalog-content\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.877240 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-utilities\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.877660 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-utilities\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.877735 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.377719879 +0000 UTC m=+153.707858739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.878167 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-catalog-content\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.896615 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4dktz"] Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.897876 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.939560 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqgzw\" (UniqueName: \"kubernetes.io/projected/1a86b0c7-517b-465f-bbe5-d957b687a38d-kube-api-access-jqgzw\") pod \"certified-operators-hjp6m\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.977089 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dktz"] Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.978984 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-catalog-content\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.979023 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph4xz\" (UniqueName: \"kubernetes.io/projected/ff5313a3-96e7-4e23-be49-8232b38d28c9-kube-api-access-ph4xz\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.979098 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-utilities\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:39 crc kubenswrapper[4679]: I1202 10:21:39.979145 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:39 crc kubenswrapper[4679]: E1202 10:21:39.979480 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.479465012 +0000 UTC m=+153.809603952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.008381 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.085052 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.085210 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph4xz\" (UniqueName: \"kubernetes.io/projected/ff5313a3-96e7-4e23-be49-8232b38d28c9-kube-api-access-ph4xz\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.085295 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-utilities\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.085386 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-catalog-content\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.085870 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-catalog-content\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.085960 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.585941777 +0000 UTC m=+153.916080647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.086485 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-utilities\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.100430 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h6jj7"] Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.101366 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.140386 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6jj7"] Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.146066 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph4xz\" (UniqueName: \"kubernetes.io/projected/ff5313a3-96e7-4e23-be49-8232b38d28c9-kube-api-access-ph4xz\") pod \"community-operators-4dktz\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.190067 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-catalog-content\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.190194 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.190233 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4r5m\" (UniqueName: \"kubernetes.io/projected/62c3d917-e199-4add-a42f-0cfdaba4dad4-kube-api-access-j4r5m\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.190265 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-utilities\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.190662 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.69064795 +0000 UTC m=+154.020786810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.278659 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.300988 4679 generic.go:334] "Generic (PLEG): container finished" podID="d40d1aca-74a1-44e7-aed1-c595ce552803" containerID="63bc391a21663da67a2d4db4ae4cf6f7d60cd5513067324208f7e9e29f0a43c8" exitCode=0 Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.301888 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" event={"ID":"d40d1aca-74a1-44e7-aed1-c595ce552803","Type":"ContainerDied","Data":"63bc391a21663da67a2d4db4ae4cf6f7d60cd5513067324208f7e9e29f0a43c8"} Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.303383 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.303766 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.311168 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.311407 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-catalog-content\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.311484 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4r5m\" (UniqueName: \"kubernetes.io/projected/62c3d917-e199-4add-a42f-0cfdaba4dad4-kube-api-access-j4r5m\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.311502 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-utilities\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.312374 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-utilities\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.312831 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-catalog-content\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.313484 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.813463112 +0000 UTC m=+154.143601982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.361355 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cjrnm" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.408395 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4r5m\" (UniqueName: \"kubernetes.io/projected/62c3d917-e199-4add-a42f-0cfdaba4dad4-kube-api-access-j4r5m\") pod \"certified-operators-h6jj7\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.433671 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.435647 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:40.935628719 +0000 UTC m=+154.265767639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.449047 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-x9sld" podStartSLOduration=10.449015597 podStartE2EDuration="10.449015597s" podCreationTimestamp="2025-12-02 10:21:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:40.448925485 +0000 UTC m=+153.779064345" watchObservedRunningTime="2025-12-02 10:21:40.449015597 +0000 UTC m=+153.779154457" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.476647 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.483176 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" podStartSLOduration=136.483154456 podStartE2EDuration="2m16.483154456s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:40.481466366 +0000 UTC m=+153.811605236" watchObservedRunningTime="2025-12-02 10:21:40.483154456 +0000 UTC m=+153.813293316" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.547661 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.548015 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.047998034 +0000 UTC m=+154.378136894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.602557 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-b689q" podStartSLOduration=136.602530087 podStartE2EDuration="2m16.602530087s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:40.565262033 +0000 UTC m=+153.895400893" watchObservedRunningTime="2025-12-02 10:21:40.602530087 +0000 UTC m=+153.932668947" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.633869 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-8jj25" podStartSLOduration=136.63385332000001 podStartE2EDuration="2m16.63385332s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:40.599641299 +0000 UTC m=+153.929780169" watchObservedRunningTime="2025-12-02 10:21:40.63385332 +0000 UTC m=+153.963992190" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.645867 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:40 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:40 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:40 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.646276 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.649347 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.649716 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.149702266 +0000 UTC m=+154.479841136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.669761 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vr2r9"] Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.751824 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.752269 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.252250328 +0000 UTC m=+154.582389188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.854885 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.855187 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.355176808 +0000 UTC m=+154.685315668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.957682 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:40 crc kubenswrapper[4679]: E1202 10:21:40.958633 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.458612481 +0000 UTC m=+154.788751341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:40 crc kubenswrapper[4679]: I1202 10:21:40.974505 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dktz"] Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.065940 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.066642 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.566622273 +0000 UTC m=+154.896761213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.182269 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.182573 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.682555152 +0000 UTC m=+155.012694012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.211184 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6jj7"] Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.267553 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hjp6m"] Dec 02 10:21:41 crc kubenswrapper[4679]: W1202 10:21:41.274717 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a86b0c7_517b_465f_bbe5_d957b687a38d.slice/crio-386ee80c7a329decc9b01889192ccda60288b164bfb2cd52f52cd45ccdaa8340 WatchSource:0}: Error finding container 386ee80c7a329decc9b01889192ccda60288b164bfb2cd52f52cd45ccdaa8340: Status 404 returned error can't find the container with id 386ee80c7a329decc9b01889192ccda60288b164bfb2cd52f52cd45ccdaa8340 Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.283339 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.283632 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.783619928 +0000 UTC m=+155.113758778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.310813 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" event={"ID":"d079b995-5285-4810-912b-0f41cde4a9d9","Type":"ContainerStarted","Data":"b3d93552de2d44457905ddb37c3bdadeacc58dbbcc561e514a7dcdf3df739c9d"} Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.315702 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6jj7" event={"ID":"62c3d917-e199-4add-a42f-0cfdaba4dad4","Type":"ContainerStarted","Data":"1861e2b8ab031f40b8d49ea73b32259858b4aeae4e70c4413bddde905d8ddaf8"} Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.317286 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerStarted","Data":"f6e43b6a587d0e6cd34d6d8ce420e7ce25babf61516fbc1bdf45c8869273c477"} Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.318356 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hjp6m" event={"ID":"1a86b0c7-517b-465f-bbe5-d957b687a38d","Type":"ContainerStarted","Data":"386ee80c7a329decc9b01889192ccda60288b164bfb2cd52f52cd45ccdaa8340"} Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.320021 4679 generic.go:334] "Generic (PLEG): container finished" podID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerID="329ba386a82361670bbabdbce0032f3226d52ffedddb89c4272179cf56aac47c" exitCode=0 Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.320824 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2r9" event={"ID":"91618844-4970-41c8-ba39-ed76f10a7ff5","Type":"ContainerDied","Data":"329ba386a82361670bbabdbce0032f3226d52ffedddb89c4272179cf56aac47c"} Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.320906 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2r9" event={"ID":"91618844-4970-41c8-ba39-ed76f10a7ff5","Type":"ContainerStarted","Data":"cd237ec8fc6e9d73195aed40baea086f5f60ca6145a329a2aed999ae9d124c6d"} Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.322656 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.384493 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.384868 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.884849279 +0000 UTC m=+155.214988149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.385585 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.385963 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.885949535 +0000 UTC m=+155.216088395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.402000 4679 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.486725 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.986702344 +0000 UTC m=+155.316841204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.487107 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.487650 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.487990 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:41.987972304 +0000 UTC m=+155.318111204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.591950 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.592129 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.092098742 +0000 UTC m=+155.422237612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.592329 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.596724 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.096705662 +0000 UTC m=+155.426844532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.638334 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:41 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:41 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:41 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.638783 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.640251 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.685938 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vmqnd"] Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.686227 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40d1aca-74a1-44e7-aed1-c595ce552803" containerName="collect-profiles" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.686241 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40d1aca-74a1-44e7-aed1-c595ce552803" containerName="collect-profiles" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.686380 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40d1aca-74a1-44e7-aed1-c595ce552803" containerName="collect-profiles" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.687273 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.691481 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693056 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq5k8\" (UniqueName: \"kubernetes.io/projected/d40d1aca-74a1-44e7-aed1-c595ce552803-kube-api-access-kq5k8\") pod \"d40d1aca-74a1-44e7-aed1-c595ce552803\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693145 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d40d1aca-74a1-44e7-aed1-c595ce552803-secret-volume\") pod \"d40d1aca-74a1-44e7-aed1-c595ce552803\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693183 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d40d1aca-74a1-44e7-aed1-c595ce552803-config-volume\") pod \"d40d1aca-74a1-44e7-aed1-c595ce552803\" (UID: \"d40d1aca-74a1-44e7-aed1-c595ce552803\") " Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693360 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693621 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-utilities\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693671 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmpmx\" (UniqueName: \"kubernetes.io/projected/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-kube-api-access-mmpmx\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.693738 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-catalog-content\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.693919 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.193905347 +0000 UTC m=+155.524044207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.694261 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40d1aca-74a1-44e7-aed1-c595ce552803-config-volume" (OuterVolumeSpecName: "config-volume") pod "d40d1aca-74a1-44e7-aed1-c595ce552803" (UID: "d40d1aca-74a1-44e7-aed1-c595ce552803"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.700681 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d40d1aca-74a1-44e7-aed1-c595ce552803-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d40d1aca-74a1-44e7-aed1-c595ce552803" (UID: "d40d1aca-74a1-44e7-aed1-c595ce552803"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.705924 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmqnd"] Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.708939 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40d1aca-74a1-44e7-aed1-c595ce552803-kube-api-access-kq5k8" (OuterVolumeSpecName: "kube-api-access-kq5k8") pod "d40d1aca-74a1-44e7-aed1-c595ce552803" (UID: "d40d1aca-74a1-44e7-aed1-c595ce552803"). InnerVolumeSpecName "kube-api-access-kq5k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.794777 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-utilities\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.794829 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmpmx\" (UniqueName: \"kubernetes.io/projected/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-kube-api-access-mmpmx\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.794877 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.794913 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-catalog-content\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.794998 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq5k8\" (UniqueName: \"kubernetes.io/projected/d40d1aca-74a1-44e7-aed1-c595ce552803-kube-api-access-kq5k8\") on node \"crc\" DevicePath \"\"" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.795015 4679 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d40d1aca-74a1-44e7-aed1-c595ce552803-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.795026 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d40d1aca-74a1-44e7-aed1-c595ce552803-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.795575 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-catalog-content\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.796110 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.29608803 +0000 UTC m=+155.626226940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.796225 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-utilities\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.813326 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmpmx\" (UniqueName: \"kubernetes.io/projected/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-kube-api-access-mmpmx\") pod \"redhat-marketplace-vmqnd\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.895880 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.896017 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.395990259 +0000 UTC m=+155.726129119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.896096 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.896439 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.396430679 +0000 UTC m=+155.726569539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.997526 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.997708 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.4976903 +0000 UTC m=+155.827829160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:41 crc kubenswrapper[4679]: I1202 10:21:41.997759 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:41 crc kubenswrapper[4679]: E1202 10:21:41.998100 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.49808481 +0000 UTC m=+155.828223670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h2ljp" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.018634 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.080782 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8w22n"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.081980 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.099769 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.099993 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-utilities\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.100088 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-catalog-content\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.100194 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjjld\" (UniqueName: \"kubernetes.io/projected/3bfee375-c85f-40ad-85fd-f16207687d6b-kube-api-access-hjjld\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: E1202 10:21:42.100423 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 10:21:42.600370915 +0000 UTC m=+155.930509795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.101476 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8w22n"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.109492 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.110776 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.112251 4679 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-02T10:21:41.402034096Z","Handler":null,"Name":""} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.117753 4679 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.117796 4679 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.118479 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.118749 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.140579 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.201244 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjjld\" (UniqueName: \"kubernetes.io/projected/3bfee375-c85f-40ad-85fd-f16207687d6b-kube-api-access-hjjld\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.201380 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.201438 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66c00c35-2791-4cca-bbdb-2615556ef599-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.201470 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-utilities\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.201489 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c00c35-2791-4cca-bbdb-2615556ef599-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.201548 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-catalog-content\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.202065 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-catalog-content\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.202945 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-utilities\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.208389 4679 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.208470 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.221794 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjjld\" (UniqueName: \"kubernetes.io/projected/3bfee375-c85f-40ad-85fd-f16207687d6b-kube-api-access-hjjld\") pod \"redhat-marketplace-8w22n\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.236618 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h2ljp\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.302278 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.302655 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66c00c35-2791-4cca-bbdb-2615556ef599-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.302688 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c00c35-2791-4cca-bbdb-2615556ef599-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.303687 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66c00c35-2791-4cca-bbdb-2615556ef599-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.315169 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.319424 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.320157 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.321109 4679 patch_prober.go:28] interesting pod/console-f9d7485db-5bk2f container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.321237 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5bk2f" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.329406 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" event={"ID":"d40d1aca-74a1-44e7-aed1-c595ce552803","Type":"ContainerDied","Data":"dd12930a522795e7837343b4f05466a7a2080a23ff7e96792e9107bc44e51748"} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.329454 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd12930a522795e7837343b4f05466a7a2080a23ff7e96792e9107bc44e51748" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.329567 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.335803 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c00c35-2791-4cca-bbdb-2615556ef599-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.344431 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.348275 4679 generic.go:334] "Generic (PLEG): container finished" podID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerID="175a9296cd7cbd6bf473b1fe30e68741c5a32a61e28e7be29a69d9176d1d8334" exitCode=0 Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.348432 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerDied","Data":"175a9296cd7cbd6bf473b1fe30e68741c5a32a61e28e7be29a69d9176d1d8334"} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.355448 4679 generic.go:334] "Generic (PLEG): container finished" podID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerID="0fc18ae0706582676d8211e94c8563222ed6528ee37e97a3f50267f62276b5fd" exitCode=0 Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.355714 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hjp6m" event={"ID":"1a86b0c7-517b-465f-bbe5-d957b687a38d","Type":"ContainerDied","Data":"0fc18ae0706582676d8211e94c8563222ed6528ee37e97a3f50267f62276b5fd"} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.362713 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" event={"ID":"d079b995-5285-4810-912b-0f41cde4a9d9","Type":"ContainerStarted","Data":"0db69f3c2d361a43729e0719457770df7fb544a66fe2f7667bd7242263b54a01"} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.362752 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" event={"ID":"d079b995-5285-4810-912b-0f41cde4a9d9","Type":"ContainerStarted","Data":"f1e6bf35cf0cf25093c1fcf273f9d9d17450f2485be254eec08efa116d0ac885"} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.368675 4679 generic.go:334] "Generic (PLEG): container finished" podID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerID="66ee60dfb30afe85c5fa4ef583569cbb15b12c26c6c6a8f2b434b71e56f65847" exitCode=0 Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.368836 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6jj7" event={"ID":"62c3d917-e199-4add-a42f-0cfdaba4dad4","Type":"ContainerDied","Data":"66ee60dfb30afe85c5fa4ef583569cbb15b12c26c6c6a8f2b434b71e56f65847"} Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.408537 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.410729 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.427526 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmqnd"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.436372 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.447354 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.448870 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.505704 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-bbxkc" podStartSLOduration=12.505687777 podStartE2EDuration="12.505687777s" podCreationTimestamp="2025-12-02 10:21:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:42.462709708 +0000 UTC m=+155.792848578" watchObservedRunningTime="2025-12-02 10:21:42.505687777 +0000 UTC m=+155.835826637" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.635828 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:42 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:42 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:42 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.635892 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.680279 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-76kbb"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.681560 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.684856 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.699102 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-76kbb"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.799654 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h2ljp"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.815114 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-catalog-content\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.815232 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-utilities\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.815272 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67pln\" (UniqueName: \"kubernetes.io/projected/dae67faf-d495-4e19-9c86-863362c49423-kube-api-access-67pln\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.915938 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-utilities\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.916011 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67pln\" (UniqueName: \"kubernetes.io/projected/dae67faf-d495-4e19-9c86-863362c49423-kube-api-access-67pln\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.916065 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-catalog-content\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.916916 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-utilities\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.917063 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-catalog-content\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.924716 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.941166 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.946909 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67pln\" (UniqueName: \"kubernetes.io/projected/dae67faf-d495-4e19-9c86-863362c49423-kube-api-access-67pln\") pod \"redhat-operators-76kbb\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.954300 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8w22n"] Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.971219 4679 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr47g container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.971288 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr47g" podUID="497dbe86-ba0e-40cb-a0ed-9608281cc6b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.971338 4679 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr47g container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Dec 02 10:21:42 crc kubenswrapper[4679]: I1202 10:21:42.971385 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zr47g" podUID="497dbe86-ba0e-40cb-a0ed-9608281cc6b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.008336 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:21:43 crc kubenswrapper[4679]: W1202 10:21:43.050257 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bfee375_c85f_40ad_85fd_f16207687d6b.slice/crio-58f9950534d6ee3eee435bd7a25bb4fd66006ef6be19a910cf25c2dc1af15635 WatchSource:0}: Error finding container 58f9950534d6ee3eee435bd7a25bb4fd66006ef6be19a910cf25c2dc1af15635: Status 404 returned error can't find the container with id 58f9950534d6ee3eee435bd7a25bb4fd66006ef6be19a910cf25c2dc1af15635 Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.075761 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jr6tk"] Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.077095 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.102835 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jr6tk"] Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.220653 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-catalog-content\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.221051 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-utilities\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.221401 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbf6j\" (UniqueName: \"kubernetes.io/projected/351b1389-f00d-4434-b58e-34a76ab674d6-kube-api-access-mbf6j\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.333490 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbf6j\" (UniqueName: \"kubernetes.io/projected/351b1389-f00d-4434-b58e-34a76ab674d6-kube-api-access-mbf6j\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.333560 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-catalog-content\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.333633 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-utilities\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.334176 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-utilities\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.335642 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-catalog-content\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.361129 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbf6j\" (UniqueName: \"kubernetes.io/projected/351b1389-f00d-4434-b58e-34a76ab674d6-kube-api-access-mbf6j\") pod \"redhat-operators-jr6tk\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.390276 4679 generic.go:334] "Generic (PLEG): container finished" podID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerID="ef0a1d61812d974c50e99536c2d5c30fe7e0c5d80be6abd4b9d26cd12fea3b35" exitCode=0 Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.391105 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8w22n" event={"ID":"3bfee375-c85f-40ad-85fd-f16207687d6b","Type":"ContainerDied","Data":"ef0a1d61812d974c50e99536c2d5c30fe7e0c5d80be6abd4b9d26cd12fea3b35"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.391130 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8w22n" event={"ID":"3bfee375-c85f-40ad-85fd-f16207687d6b","Type":"ContainerStarted","Data":"58f9950534d6ee3eee435bd7a25bb4fd66006ef6be19a910cf25c2dc1af15635"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.395802 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"66c00c35-2791-4cca-bbdb-2615556ef599","Type":"ContainerStarted","Data":"71456229ae0bc86f4e2dda4158a7af0109da3507c9a1a6fe552bf4770fb82675"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.398245 4679 generic.go:334] "Generic (PLEG): container finished" podID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerID="fd878ff823515af7ce151a0631d27abdb64dc7abe8e5368211735726fce21233" exitCode=0 Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.399257 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmqnd" event={"ID":"3dfcd027-05cf-48bc-8f61-f14a8d2a611f","Type":"ContainerDied","Data":"fd878ff823515af7ce151a0631d27abdb64dc7abe8e5368211735726fce21233"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.399323 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmqnd" event={"ID":"3dfcd027-05cf-48bc-8f61-f14a8d2a611f","Type":"ContainerStarted","Data":"5158e51d1e26c148ddcd37fb389a43d8ae12a3ab9db27bba0544f42cfcb1ef45"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.400280 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.404280 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" event={"ID":"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26","Type":"ContainerStarted","Data":"7bb13bdcf48b735fe12f95e6ba5a297afa05925396a7c2f6b96ddb17d137e050"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.404345 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" event={"ID":"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26","Type":"ContainerStarted","Data":"d1d6721be96aa935f23fee51dd81fa8a1bbf9eed3883ac7b1491235abb5d556d"} Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.419526 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-s6wzr" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.459821 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" podStartSLOduration=139.459800662 podStartE2EDuration="2m19.459800662s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:21:43.458156133 +0000 UTC m=+156.788294993" watchObservedRunningTime="2025-12-02 10:21:43.459800662 +0000 UTC m=+156.789939532" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.634699 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.638214 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:43 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:43 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:43 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.638263 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.650112 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-76kbb"] Dec 02 10:21:43 crc kubenswrapper[4679]: W1202 10:21:43.673188 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddae67faf_d495_4e19_9c86_863362c49423.slice/crio-5afe77130a795005eb328cd71825c9571d8420a7f908ed6834785f104c15d398 WatchSource:0}: Error finding container 5afe77130a795005eb328cd71825c9571d8420a7f908ed6834785f104c15d398: Status 404 returned error can't find the container with id 5afe77130a795005eb328cd71825c9571d8420a7f908ed6834785f104c15d398 Dec 02 10:21:43 crc kubenswrapper[4679]: I1202 10:21:43.775859 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jr6tk"] Dec 02 10:21:43 crc kubenswrapper[4679]: W1202 10:21:43.784528 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod351b1389_f00d_4434_b58e_34a76ab674d6.slice/crio-ad3aa2f18dc11344862e2deb61019573fec37fbcc9d418af585c1d1bb2de7614 WatchSource:0}: Error finding container ad3aa2f18dc11344862e2deb61019573fec37fbcc9d418af585c1d1bb2de7614: Status 404 returned error can't find the container with id ad3aa2f18dc11344862e2deb61019573fec37fbcc9d418af585c1d1bb2de7614 Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.099572 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.102589 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.414877 4679 generic.go:334] "Generic (PLEG): container finished" podID="dae67faf-d495-4e19-9c86-863362c49423" containerID="c9e1947fb9faf037921636b1447729913f8722ca7e188581921020c1f2e750d6" exitCode=0 Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.414969 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerDied","Data":"c9e1947fb9faf037921636b1447729913f8722ca7e188581921020c1f2e750d6"} Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.415003 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerStarted","Data":"5afe77130a795005eb328cd71825c9571d8420a7f908ed6834785f104c15d398"} Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.422480 4679 generic.go:334] "Generic (PLEG): container finished" podID="351b1389-f00d-4434-b58e-34a76ab674d6" containerID="3b6634a93fb85a66ed330f0e60b05da30f42b3910ad4872424f5a6f9fc0d716d" exitCode=0 Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.422559 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerDied","Data":"3b6634a93fb85a66ed330f0e60b05da30f42b3910ad4872424f5a6f9fc0d716d"} Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.422591 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerStarted","Data":"ad3aa2f18dc11344862e2deb61019573fec37fbcc9d418af585c1d1bb2de7614"} Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.426004 4679 generic.go:334] "Generic (PLEG): container finished" podID="66c00c35-2791-4cca-bbdb-2615556ef599" containerID="34032fe5d8a5353e498b2674d2f8185ce732b280d0a627924e906a8f2234b479" exitCode=0 Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.426728 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"66c00c35-2791-4cca-bbdb-2615556ef599","Type":"ContainerDied","Data":"34032fe5d8a5353e498b2674d2f8185ce732b280d0a627924e906a8f2234b479"} Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.426760 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.639493 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:44 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:44 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:44 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:44 crc kubenswrapper[4679]: I1202 10:21:44.639589 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:45 crc kubenswrapper[4679]: I1202 10:21:45.643576 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:45 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:45 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:45 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:45 crc kubenswrapper[4679]: I1202 10:21:45.643634 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:45 crc kubenswrapper[4679]: I1202 10:21:45.930825 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:45 crc kubenswrapper[4679]: I1202 10:21:45.990918 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66c00c35-2791-4cca-bbdb-2615556ef599-kubelet-dir\") pod \"66c00c35-2791-4cca-bbdb-2615556ef599\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " Dec 02 10:21:45 crc kubenswrapper[4679]: I1202 10:21:45.991078 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c00c35-2791-4cca-bbdb-2615556ef599-kube-api-access\") pod \"66c00c35-2791-4cca-bbdb-2615556ef599\" (UID: \"66c00c35-2791-4cca-bbdb-2615556ef599\") " Dec 02 10:21:45 crc kubenswrapper[4679]: I1202 10:21:45.991390 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66c00c35-2791-4cca-bbdb-2615556ef599-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "66c00c35-2791-4cca-bbdb-2615556ef599" (UID: "66c00c35-2791-4cca-bbdb-2615556ef599"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.022981 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c00c35-2791-4cca-bbdb-2615556ef599-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "66c00c35-2791-4cca-bbdb-2615556ef599" (UID: "66c00c35-2791-4cca-bbdb-2615556ef599"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.092571 4679 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66c00c35-2791-4cca-bbdb-2615556ef599-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.092600 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c00c35-2791-4cca-bbdb-2615556ef599-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.452052 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"66c00c35-2791-4cca-bbdb-2615556ef599","Type":"ContainerDied","Data":"71456229ae0bc86f4e2dda4158a7af0109da3507c9a1a6fe552bf4770fb82675"} Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.452131 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71456229ae0bc86f4e2dda4158a7af0109da3507c9a1a6fe552bf4770fb82675" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.452184 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.636109 4679 patch_prober.go:28] interesting pod/router-default-5444994796-bf564 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 10:21:46 crc kubenswrapper[4679]: [-]has-synced failed: reason withheld Dec 02 10:21:46 crc kubenswrapper[4679]: [+]process-running ok Dec 02 10:21:46 crc kubenswrapper[4679]: healthz check failed Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.636187 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bf564" podUID="0f5b6f36-f737-42b9-9403-db45098c682a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.931813 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:21:46 crc kubenswrapper[4679]: I1202 10:21:46.931879 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.112616 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.117540 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1827954-0852-40f4-ad94-c979f72addb9-metrics-certs\") pod \"network-metrics-daemon-xlg8w\" (UID: \"c1827954-0852-40f4-ad94-c979f72addb9\") " pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.244720 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xlg8w" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.307198 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 10:21:47 crc kubenswrapper[4679]: E1202 10:21:47.307492 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c00c35-2791-4cca-bbdb-2615556ef599" containerName="pruner" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.307508 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c00c35-2791-4cca-bbdb-2615556ef599" containerName="pruner" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.307641 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c00c35-2791-4cca-bbdb-2615556ef599" containerName="pruner" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.308131 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.309896 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.310352 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.315200 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.419918 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fd050d9-5ca0-4439-a758-ae85e28f334e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.420139 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6fd050d9-5ca0-4439-a758-ae85e28f334e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.476030 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.521525 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6fd050d9-5ca0-4439-a758-ae85e28f334e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.521606 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fd050d9-5ca0-4439-a758-ae85e28f334e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.522003 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6fd050d9-5ca0-4439-a758-ae85e28f334e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.547131 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fd050d9-5ca0-4439-a758-ae85e28f334e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.640600 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.648033 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:47 crc kubenswrapper[4679]: I1202 10:21:47.663675 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bf564" Dec 02 10:21:49 crc kubenswrapper[4679]: I1202 10:21:49.159627 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-x9sld" Dec 02 10:21:52 crc kubenswrapper[4679]: I1202 10:21:52.319676 4679 patch_prober.go:28] interesting pod/console-f9d7485db-5bk2f container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 02 10:21:52 crc kubenswrapper[4679]: I1202 10:21:52.320108 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5bk2f" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 02 10:21:52 crc kubenswrapper[4679]: I1202 10:21:52.980561 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-zr47g" Dec 02 10:22:02 crc kubenswrapper[4679]: I1202 10:22:02.331161 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:22:02 crc kubenswrapper[4679]: I1202 10:22:02.335388 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:22:02 crc kubenswrapper[4679]: I1202 10:22:02.349917 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:22:13 crc kubenswrapper[4679]: I1202 10:22:13.157958 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 10:22:14 crc kubenswrapper[4679]: I1202 10:22:14.064685 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9fzfq" Dec 02 10:22:16 crc kubenswrapper[4679]: I1202 10:22:16.932028 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:22:16 crc kubenswrapper[4679]: I1202 10:22:16.932421 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:22:18 crc kubenswrapper[4679]: E1202 10:22:18.522457 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 02 10:22:18 crc kubenswrapper[4679]: E1202 10:22:18.522947 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jqgzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hjp6m_openshift-marketplace(1a86b0c7-517b-465f-bbe5-d957b687a38d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:18 crc kubenswrapper[4679]: E1202 10:22:18.524155 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hjp6m" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" Dec 02 10:22:18 crc kubenswrapper[4679]: E1202 10:22:18.531197 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 02 10:22:18 crc kubenswrapper[4679]: E1202 10:22:18.531368 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j4r5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-h6jj7_openshift-marketplace(62c3d917-e199-4add-a42f-0cfdaba4dad4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:18 crc kubenswrapper[4679]: E1202 10:22:18.532502 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-h6jj7" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.205473 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-h6jj7" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.205580 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hjp6m" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.379489 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.379636 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hjjld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8w22n_openshift-marketplace(3bfee375-c85f-40ad-85fd-f16207687d6b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.380836 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8w22n" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.592093 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.592278 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mmpmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vmqnd_openshift-marketplace(3dfcd027-05cf-48bc-8f61-f14a8d2a611f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:20 crc kubenswrapper[4679]: E1202 10:22:20.593462 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vmqnd" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" Dec 02 10:22:21 crc kubenswrapper[4679]: I1202 10:22:21.880555 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 10:22:21 crc kubenswrapper[4679]: I1202 10:22:21.884532 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:21 crc kubenswrapper[4679]: I1202 10:22:21.904591 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.909919 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vmqnd" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.910223 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8w22n" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.976847 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.977021 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ph4xz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4dktz_openshift-marketplace(ff5313a3-96e7-4e23-be49-8232b38d28c9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.978156 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4dktz" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" Dec 02 10:22:21 crc kubenswrapper[4679]: I1202 10:22:21.990611 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:21 crc kubenswrapper[4679]: I1202 10:22:21.990690 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.996385 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.996518 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99btx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vr2r9_openshift-marketplace(91618844-4970-41c8-ba39-ed76f10a7ff5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:21 crc kubenswrapper[4679]: E1202 10:22:21.997986 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vr2r9" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" Dec 02 10:22:22 crc kubenswrapper[4679]: I1202 10:22:22.092139 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:22 crc kubenswrapper[4679]: I1202 10:22:22.092214 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:22 crc kubenswrapper[4679]: I1202 10:22:22.092280 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:22 crc kubenswrapper[4679]: I1202 10:22:22.113741 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:22 crc kubenswrapper[4679]: I1202 10:22:22.219007 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:25 crc kubenswrapper[4679]: E1202 10:22:25.052725 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vr2r9" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" Dec 02 10:22:25 crc kubenswrapper[4679]: E1202 10:22:25.052743 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4dktz" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" Dec 02 10:22:25 crc kubenswrapper[4679]: E1202 10:22:25.084166 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 02 10:22:25 crc kubenswrapper[4679]: E1202 10:22:25.085153 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mbf6j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jr6tk_openshift-marketplace(351b1389-f00d-4434-b58e-34a76ab674d6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 10:22:25 crc kubenswrapper[4679]: E1202 10:22:25.086352 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jr6tk" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.452653 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.463258 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xlg8w"] Dec 02 10:22:25 crc kubenswrapper[4679]: W1202 10:22:25.466951 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6fd050d9_5ca0_4439_a758_ae85e28f334e.slice/crio-439c7de1666d77ff162fe698ff325d556f385e4372e5702cec32136ced3f2aa6 WatchSource:0}: Error finding container 439c7de1666d77ff162fe698ff325d556f385e4372e5702cec32136ced3f2aa6: Status 404 returned error can't find the container with id 439c7de1666d77ff162fe698ff325d556f385e4372e5702cec32136ced3f2aa6 Dec 02 10:22:25 crc kubenswrapper[4679]: W1202 10:22:25.471801 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1827954_0852_40f4_ad94_c979f72addb9.slice/crio-275d74ea3cefaa490a9db077b57accab8168c68299da48a7f8182e8cc6693dd3 WatchSource:0}: Error finding container 275d74ea3cefaa490a9db077b57accab8168c68299da48a7f8182e8cc6693dd3: Status 404 returned error can't find the container with id 275d74ea3cefaa490a9db077b57accab8168c68299da48a7f8182e8cc6693dd3 Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.593879 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 10:22:25 crc kubenswrapper[4679]: W1202 10:22:25.606072 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod614d686a_6f9a_47c2_93f0_0a6722e0d3fb.slice/crio-594924b1fbc99d2a5bf429beaeac4e2948a21796fb947d6d65a298e75f8618f1 WatchSource:0}: Error finding container 594924b1fbc99d2a5bf429beaeac4e2948a21796fb947d6d65a298e75f8618f1: Status 404 returned error can't find the container with id 594924b1fbc99d2a5bf429beaeac4e2948a21796fb947d6d65a298e75f8618f1 Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.771342 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6fd050d9-5ca0-4439-a758-ae85e28f334e","Type":"ContainerStarted","Data":"439c7de1666d77ff162fe698ff325d556f385e4372e5702cec32136ced3f2aa6"} Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.772518 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" event={"ID":"c1827954-0852-40f4-ad94-c979f72addb9","Type":"ContainerStarted","Data":"571c67bdaba98346e44f366b90d389fbb7390086d32662f37604d897850ec7da"} Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.772550 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" event={"ID":"c1827954-0852-40f4-ad94-c979f72addb9","Type":"ContainerStarted","Data":"275d74ea3cefaa490a9db077b57accab8168c68299da48a7f8182e8cc6693dd3"} Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.776729 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerStarted","Data":"7c9695d9473b5f50bbec8b1ae164aca801823e2dca043be1ca2c064d3d41aef8"} Dec 02 10:22:25 crc kubenswrapper[4679]: I1202 10:22:25.778285 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"614d686a-6f9a-47c2-93f0-0a6722e0d3fb","Type":"ContainerStarted","Data":"594924b1fbc99d2a5bf429beaeac4e2948a21796fb947d6d65a298e75f8618f1"} Dec 02 10:22:25 crc kubenswrapper[4679]: E1202 10:22:25.780612 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jr6tk" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.784568 4679 generic.go:334] "Generic (PLEG): container finished" podID="6fd050d9-5ca0-4439-a758-ae85e28f334e" containerID="d57bf782d3e6fab3f7269c77eb21fc26abff177212fde5f34fe2aa79516d272c" exitCode=0 Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.784658 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6fd050d9-5ca0-4439-a758-ae85e28f334e","Type":"ContainerDied","Data":"d57bf782d3e6fab3f7269c77eb21fc26abff177212fde5f34fe2aa79516d272c"} Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.788229 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xlg8w" event={"ID":"c1827954-0852-40f4-ad94-c979f72addb9","Type":"ContainerStarted","Data":"a6722835b8f5d6f2f2292d3b32aef53c5a8d648fe93b97728d1004c0ee01f69f"} Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.790556 4679 generic.go:334] "Generic (PLEG): container finished" podID="dae67faf-d495-4e19-9c86-863362c49423" containerID="7c9695d9473b5f50bbec8b1ae164aca801823e2dca043be1ca2c064d3d41aef8" exitCode=0 Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.790610 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerDied","Data":"7c9695d9473b5f50bbec8b1ae164aca801823e2dca043be1ca2c064d3d41aef8"} Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.799822 4679 generic.go:334] "Generic (PLEG): container finished" podID="614d686a-6f9a-47c2-93f0-0a6722e0d3fb" containerID="9c96b8caf10cf078991599f434da50d46c6f16e9db03ccedf282ba6cb8f7434f" exitCode=0 Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.799892 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"614d686a-6f9a-47c2-93f0-0a6722e0d3fb","Type":"ContainerDied","Data":"9c96b8caf10cf078991599f434da50d46c6f16e9db03ccedf282ba6cb8f7434f"} Dec 02 10:22:26 crc kubenswrapper[4679]: I1202 10:22:26.834734 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xlg8w" podStartSLOduration=182.83471089 podStartE2EDuration="3m2.83471089s" podCreationTimestamp="2025-12-02 10:19:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:22:26.827943525 +0000 UTC m=+200.158082385" watchObservedRunningTime="2025-12-02 10:22:26.83471089 +0000 UTC m=+200.164849760" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.077047 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.077710 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.094417 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.256001 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-var-lock\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.256104 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.256275 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36424a7-1339-47b3-918a-4c51721b3476-kube-api-access\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.357292 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-var-lock\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.357636 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.357715 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.357369 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-var-lock\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.357953 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36424a7-1339-47b3-918a-4c51721b3476-kube-api-access\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.377111 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36424a7-1339-47b3-918a-4c51721b3476-kube-api-access\") pod \"installer-9-crc\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:27 crc kubenswrapper[4679]: I1202 10:22:27.404899 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:22:28 crc kubenswrapper[4679]: I1202 10:22:28.922765 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:28 crc kubenswrapper[4679]: I1202 10:22:28.926619 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.044049 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 10:22:29 crc kubenswrapper[4679]: W1202 10:22:29.065862 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda36424a7_1339_47b3_918a_4c51721b3476.slice/crio-1d5a73b998338561c344b2cad9dc7ee130184029073dcc981821a5556dce29d2 WatchSource:0}: Error finding container 1d5a73b998338561c344b2cad9dc7ee130184029073dcc981821a5556dce29d2: Status 404 returned error can't find the container with id 1d5a73b998338561c344b2cad9dc7ee130184029073dcc981821a5556dce29d2 Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.083944 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fd050d9-5ca0-4439-a758-ae85e28f334e-kube-api-access\") pod \"6fd050d9-5ca0-4439-a758-ae85e28f334e\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.085412 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kube-api-access\") pod \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.085447 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kubelet-dir\") pod \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\" (UID: \"614d686a-6f9a-47c2-93f0-0a6722e0d3fb\") " Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.085527 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6fd050d9-5ca0-4439-a758-ae85e28f334e-kubelet-dir\") pod \"6fd050d9-5ca0-4439-a758-ae85e28f334e\" (UID: \"6fd050d9-5ca0-4439-a758-ae85e28f334e\") " Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.085639 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "614d686a-6f9a-47c2-93f0-0a6722e0d3fb" (UID: "614d686a-6f9a-47c2-93f0-0a6722e0d3fb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.085699 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6fd050d9-5ca0-4439-a758-ae85e28f334e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6fd050d9-5ca0-4439-a758-ae85e28f334e" (UID: "6fd050d9-5ca0-4439-a758-ae85e28f334e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.085975 4679 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.086001 4679 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6fd050d9-5ca0-4439-a758-ae85e28f334e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.092627 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "614d686a-6f9a-47c2-93f0-0a6722e0d3fb" (UID: "614d686a-6f9a-47c2-93f0-0a6722e0d3fb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.096577 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fd050d9-5ca0-4439-a758-ae85e28f334e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6fd050d9-5ca0-4439-a758-ae85e28f334e" (UID: "6fd050d9-5ca0-4439-a758-ae85e28f334e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.187528 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fd050d9-5ca0-4439-a758-ae85e28f334e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.187572 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614d686a-6f9a-47c2-93f0-0a6722e0d3fb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.816108 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a36424a7-1339-47b3-918a-4c51721b3476","Type":"ContainerStarted","Data":"4161ba6c8d8a4479c07c51fbdef4e842b214ffe4b87e349edd75a8ca16ac73a2"} Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.816496 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a36424a7-1339-47b3-918a-4c51721b3476","Type":"ContainerStarted","Data":"1d5a73b998338561c344b2cad9dc7ee130184029073dcc981821a5556dce29d2"} Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.817791 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.817822 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"6fd050d9-5ca0-4439-a758-ae85e28f334e","Type":"ContainerDied","Data":"439c7de1666d77ff162fe698ff325d556f385e4372e5702cec32136ced3f2aa6"} Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.817852 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="439c7de1666d77ff162fe698ff325d556f385e4372e5702cec32136ced3f2aa6" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.824002 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerStarted","Data":"3eb5d76f591f2e22cd217e976963c9c4bf13cb4dd04f3d14b87cdf2441e3b5b0"} Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.825414 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"614d686a-6f9a-47c2-93f0-0a6722e0d3fb","Type":"ContainerDied","Data":"594924b1fbc99d2a5bf429beaeac4e2948a21796fb947d6d65a298e75f8618f1"} Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.825444 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="594924b1fbc99d2a5bf429beaeac4e2948a21796fb947d6d65a298e75f8618f1" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.825475 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.835807 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.835779225 podStartE2EDuration="2.835779225s" podCreationTimestamp="2025-12-02 10:22:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:22:29.827401833 +0000 UTC m=+203.157540713" watchObservedRunningTime="2025-12-02 10:22:29.835779225 +0000 UTC m=+203.165918085" Dec 02 10:22:29 crc kubenswrapper[4679]: I1202 10:22:29.852190 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-76kbb" podStartSLOduration=3.630536156 podStartE2EDuration="47.852173068s" podCreationTimestamp="2025-12-02 10:21:42 +0000 UTC" firstStartedPulling="2025-12-02 10:21:44.418760852 +0000 UTC m=+157.748899712" lastFinishedPulling="2025-12-02 10:22:28.640397754 +0000 UTC m=+201.970536624" observedRunningTime="2025-12-02 10:22:29.846583407 +0000 UTC m=+203.176722277" watchObservedRunningTime="2025-12-02 10:22:29.852173068 +0000 UTC m=+203.182311928" Dec 02 10:22:32 crc kubenswrapper[4679]: I1202 10:22:32.848639 4679 generic.go:334] "Generic (PLEG): container finished" podID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerID="25a3bc0a7188d284f8476625c3ea947b515951f5e89d61ba63097b753b98fa95" exitCode=0 Dec 02 10:22:32 crc kubenswrapper[4679]: I1202 10:22:32.848678 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hjp6m" event={"ID":"1a86b0c7-517b-465f-bbe5-d957b687a38d","Type":"ContainerDied","Data":"25a3bc0a7188d284f8476625c3ea947b515951f5e89d61ba63097b753b98fa95"} Dec 02 10:22:33 crc kubenswrapper[4679]: I1202 10:22:33.009504 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:22:33 crc kubenswrapper[4679]: I1202 10:22:33.009575 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:22:33 crc kubenswrapper[4679]: I1202 10:22:33.860634 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hjp6m" event={"ID":"1a86b0c7-517b-465f-bbe5-d957b687a38d","Type":"ContainerStarted","Data":"7fb5a702b248b59c0339c642ffd4fa888329f5b181d4402c2e1f041cb7c47ad8"} Dec 02 10:22:33 crc kubenswrapper[4679]: I1202 10:22:33.881894 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hjp6m" podStartSLOduration=3.886519895 podStartE2EDuration="54.881875884s" podCreationTimestamp="2025-12-02 10:21:39 +0000 UTC" firstStartedPulling="2025-12-02 10:21:42.356887668 +0000 UTC m=+155.687026528" lastFinishedPulling="2025-12-02 10:22:33.352243647 +0000 UTC m=+206.682382517" observedRunningTime="2025-12-02 10:22:33.878290121 +0000 UTC m=+207.208428991" watchObservedRunningTime="2025-12-02 10:22:33.881875884 +0000 UTC m=+207.212014744" Dec 02 10:22:34 crc kubenswrapper[4679]: I1202 10:22:34.063621 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-76kbb" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="registry-server" probeResult="failure" output=< Dec 02 10:22:34 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 10:22:34 crc kubenswrapper[4679]: > Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.891880 4679 generic.go:334] "Generic (PLEG): container finished" podID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerID="0185b16bd01dcc02ae78ff581bd5ed8f87285d71372f47355ed678a8ee317f73" exitCode=0 Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.891960 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8w22n" event={"ID":"3bfee375-c85f-40ad-85fd-f16207687d6b","Type":"ContainerDied","Data":"0185b16bd01dcc02ae78ff581bd5ed8f87285d71372f47355ed678a8ee317f73"} Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.896155 4679 generic.go:334] "Generic (PLEG): container finished" podID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerID="9552f093fa92daf277a9923c688c4b9f472b8ec7876e72dbf12e314c40965f03" exitCode=0 Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.896224 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmqnd" event={"ID":"3dfcd027-05cf-48bc-8f61-f14a8d2a611f","Type":"ContainerDied","Data":"9552f093fa92daf277a9923c688c4b9f472b8ec7876e72dbf12e314c40965f03"} Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.898039 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerStarted","Data":"920e3d597e47adaff46a1622ce350d25d8faf577e7b3a19dc06287fda8412873"} Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.900131 4679 generic.go:334] "Generic (PLEG): container finished" podID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerID="a3b8b36118f67bd6af2327e1ab657d5c9902c0c0b79bca5f7fdf1307e463ab00" exitCode=0 Dec 02 10:22:39 crc kubenswrapper[4679]: I1202 10:22:39.900174 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6jj7" event={"ID":"62c3d917-e199-4add-a42f-0cfdaba4dad4","Type":"ContainerDied","Data":"a3b8b36118f67bd6af2327e1ab657d5c9902c0c0b79bca5f7fdf1307e463ab00"} Dec 02 10:22:40 crc kubenswrapper[4679]: I1202 10:22:40.008908 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:22:40 crc kubenswrapper[4679]: I1202 10:22:40.008996 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:22:40 crc kubenswrapper[4679]: I1202 10:22:40.188006 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:22:40 crc kubenswrapper[4679]: I1202 10:22:40.906187 4679 generic.go:334] "Generic (PLEG): container finished" podID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerID="920e3d597e47adaff46a1622ce350d25d8faf577e7b3a19dc06287fda8412873" exitCode=0 Dec 02 10:22:40 crc kubenswrapper[4679]: I1202 10:22:40.906290 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerDied","Data":"920e3d597e47adaff46a1622ce350d25d8faf577e7b3a19dc06287fda8412873"} Dec 02 10:22:40 crc kubenswrapper[4679]: I1202 10:22:40.955032 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:22:43 crc kubenswrapper[4679]: I1202 10:22:43.055509 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:22:43 crc kubenswrapper[4679]: I1202 10:22:43.098974 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:22:46 crc kubenswrapper[4679]: I1202 10:22:46.931275 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:22:46 crc kubenswrapper[4679]: I1202 10:22:46.931528 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:22:46 crc kubenswrapper[4679]: I1202 10:22:46.931574 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:22:46 crc kubenswrapper[4679]: I1202 10:22:46.934857 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:22:46 crc kubenswrapper[4679]: I1202 10:22:46.934986 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311" gracePeriod=600 Dec 02 10:22:48 crc kubenswrapper[4679]: I1202 10:22:48.944257 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311" exitCode=0 Dec 02 10:22:48 crc kubenswrapper[4679]: I1202 10:22:48.944367 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.958202 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerStarted","Data":"e4f0ba02e8f6356b3c1439b038f89ff5b5ebf9c5f9bc75ee9d132627626fd61f"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.961655 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"82a102b5f7c254ea6255ad947fef5a434fb11ff49e4dbe4618063250f135f38a"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.963731 4679 generic.go:334] "Generic (PLEG): container finished" podID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerID="d804594ed1b5f3aba71e2532c9a79b1fb58a80df90f86d872d08313ac87c2d93" exitCode=0 Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.963797 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2r9" event={"ID":"91618844-4970-41c8-ba39-ed76f10a7ff5","Type":"ContainerDied","Data":"d804594ed1b5f3aba71e2532c9a79b1fb58a80df90f86d872d08313ac87c2d93"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.966147 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6jj7" event={"ID":"62c3d917-e199-4add-a42f-0cfdaba4dad4","Type":"ContainerStarted","Data":"69c6bfab086d5136a7ef6ce7b5f85730936be6d8960dd26d3436953a479e822c"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.970461 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8w22n" event={"ID":"3bfee375-c85f-40ad-85fd-f16207687d6b","Type":"ContainerStarted","Data":"a2cb47d9babbb438804391ffc71d92c6ee6812ef9865bb25a288e8c58a86dc08"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.971942 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerStarted","Data":"2936e1212b4ace676c019115c2bcbe9a0e5022c52da16dcd6ea2e1a0e1b21832"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.974624 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmqnd" event={"ID":"3dfcd027-05cf-48bc-8f61-f14a8d2a611f","Type":"ContainerStarted","Data":"4b65718c7eb0d676bdea3bacafa0759ee0a10fc60ef7f082a16d2ec2b85a0f9b"} Dec 02 10:22:50 crc kubenswrapper[4679]: I1202 10:22:50.987358 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4dktz" podStartSLOduration=4.449531471 podStartE2EDuration="1m11.987339527s" podCreationTimestamp="2025-12-02 10:21:39 +0000 UTC" firstStartedPulling="2025-12-02 10:21:42.353927928 +0000 UTC m=+155.684066788" lastFinishedPulling="2025-12-02 10:22:49.891735984 +0000 UTC m=+223.221874844" observedRunningTime="2025-12-02 10:22:50.985532035 +0000 UTC m=+224.315670895" watchObservedRunningTime="2025-12-02 10:22:50.987339527 +0000 UTC m=+224.317478387" Dec 02 10:22:51 crc kubenswrapper[4679]: I1202 10:22:51.008448 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h6jj7" podStartSLOduration=3.532073931 podStartE2EDuration="1m11.008427855s" podCreationTimestamp="2025-12-02 10:21:40 +0000 UTC" firstStartedPulling="2025-12-02 10:21:42.39701055 +0000 UTC m=+155.727149420" lastFinishedPulling="2025-12-02 10:22:49.873364484 +0000 UTC m=+223.203503344" observedRunningTime="2025-12-02 10:22:51.004605425 +0000 UTC m=+224.334744285" watchObservedRunningTime="2025-12-02 10:22:51.008427855 +0000 UTC m=+224.338566715" Dec 02 10:22:51 crc kubenswrapper[4679]: I1202 10:22:51.026134 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8w22n" podStartSLOduration=2.548403444 podStartE2EDuration="1m9.026107735s" podCreationTimestamp="2025-12-02 10:21:42 +0000 UTC" firstStartedPulling="2025-12-02 10:21:43.395347034 +0000 UTC m=+156.725485894" lastFinishedPulling="2025-12-02 10:22:49.873051305 +0000 UTC m=+223.203190185" observedRunningTime="2025-12-02 10:22:51.02558299 +0000 UTC m=+224.355721850" watchObservedRunningTime="2025-12-02 10:22:51.026107735 +0000 UTC m=+224.356246595" Dec 02 10:22:51 crc kubenswrapper[4679]: I1202 10:22:51.165688 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vmqnd" podStartSLOduration=3.6921470469999997 podStartE2EDuration="1m10.165667551s" podCreationTimestamp="2025-12-02 10:21:41 +0000 UTC" firstStartedPulling="2025-12-02 10:21:43.39982699 +0000 UTC m=+156.729965850" lastFinishedPulling="2025-12-02 10:22:49.873347484 +0000 UTC m=+223.203486354" observedRunningTime="2025-12-02 10:22:51.143942464 +0000 UTC m=+224.474081324" watchObservedRunningTime="2025-12-02 10:22:51.165667551 +0000 UTC m=+224.495806411" Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.019776 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.020040 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.275062 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-68gpm"] Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.437735 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.437967 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.984832 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2r9" event={"ID":"91618844-4970-41c8-ba39-ed76f10a7ff5","Type":"ContainerStarted","Data":"89171d85207d29fb63658d2acd7ae4d962339fb3b9484fc692fc136200ad18f9"} Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.987561 4679 generic.go:334] "Generic (PLEG): container finished" podID="351b1389-f00d-4434-b58e-34a76ab674d6" containerID="2936e1212b4ace676c019115c2bcbe9a0e5022c52da16dcd6ea2e1a0e1b21832" exitCode=0 Dec 02 10:22:52 crc kubenswrapper[4679]: I1202 10:22:52.987593 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerDied","Data":"2936e1212b4ace676c019115c2bcbe9a0e5022c52da16dcd6ea2e1a0e1b21832"} Dec 02 10:22:53 crc kubenswrapper[4679]: I1202 10:22:53.005847 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vr2r9" podStartSLOduration=3.580229991 podStartE2EDuration="1m14.005827831s" podCreationTimestamp="2025-12-02 10:21:39 +0000 UTC" firstStartedPulling="2025-12-02 10:21:41.322283795 +0000 UTC m=+154.652422655" lastFinishedPulling="2025-12-02 10:22:51.747881635 +0000 UTC m=+225.078020495" observedRunningTime="2025-12-02 10:22:53.002039221 +0000 UTC m=+226.332178091" watchObservedRunningTime="2025-12-02 10:22:53.005827831 +0000 UTC m=+226.335966701" Dec 02 10:22:53 crc kubenswrapper[4679]: I1202 10:22:53.097246 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-vmqnd" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="registry-server" probeResult="failure" output=< Dec 02 10:22:53 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 10:22:53 crc kubenswrapper[4679]: > Dec 02 10:22:53 crc kubenswrapper[4679]: I1202 10:22:53.590507 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-8w22n" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="registry-server" probeResult="failure" output=< Dec 02 10:22:53 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 10:22:53 crc kubenswrapper[4679]: > Dec 02 10:22:55 crc kubenswrapper[4679]: I1202 10:22:54.999971 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerStarted","Data":"46dbd29969ab03b89fbd2b6a4d4cbb3d5d826f0a2c441428af64cb8044aedca5"} Dec 02 10:22:55 crc kubenswrapper[4679]: I1202 10:22:55.024443 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jr6tk" podStartSLOduration=1.85781671 podStartE2EDuration="1m12.024409556s" podCreationTimestamp="2025-12-02 10:21:43 +0000 UTC" firstStartedPulling="2025-12-02 10:21:44.431912014 +0000 UTC m=+157.762050874" lastFinishedPulling="2025-12-02 10:22:54.59850486 +0000 UTC m=+227.928643720" observedRunningTime="2025-12-02 10:22:55.016906389 +0000 UTC m=+228.347045269" watchObservedRunningTime="2025-12-02 10:22:55.024409556 +0000 UTC m=+228.354548456" Dec 02 10:22:59 crc kubenswrapper[4679]: I1202 10:22:59.807955 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:22:59 crc kubenswrapper[4679]: I1202 10:22:59.808519 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:22:59 crc kubenswrapper[4679]: I1202 10:22:59.854563 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.072198 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.279707 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.279768 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.347930 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.477553 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.477595 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:23:00 crc kubenswrapper[4679]: I1202 10:23:00.526433 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:23:01 crc kubenswrapper[4679]: I1202 10:23:01.086609 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:23:01 crc kubenswrapper[4679]: I1202 10:23:01.125870 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:23:01 crc kubenswrapper[4679]: I1202 10:23:01.891429 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dktz"] Dec 02 10:23:02 crc kubenswrapper[4679]: I1202 10:23:02.061027 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:23:02 crc kubenswrapper[4679]: I1202 10:23:02.108224 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:23:02 crc kubenswrapper[4679]: I1202 10:23:02.482633 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:23:02 crc kubenswrapper[4679]: I1202 10:23:02.524819 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:23:02 crc kubenswrapper[4679]: I1202 10:23:02.803228 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h6jj7"] Dec 02 10:23:03 crc kubenswrapper[4679]: I1202 10:23:03.054691 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h6jj7" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="registry-server" containerID="cri-o://69c6bfab086d5136a7ef6ce7b5f85730936be6d8960dd26d3436953a479e822c" gracePeriod=2 Dec 02 10:23:03 crc kubenswrapper[4679]: I1202 10:23:03.054855 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4dktz" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="registry-server" containerID="cri-o://e4f0ba02e8f6356b3c1439b038f89ff5b5ebf9c5f9bc75ee9d132627626fd61f" gracePeriod=2 Dec 02 10:23:03 crc kubenswrapper[4679]: I1202 10:23:03.401707 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:23:03 crc kubenswrapper[4679]: I1202 10:23:03.401796 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:23:03 crc kubenswrapper[4679]: I1202 10:23:03.440945 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:23:04 crc kubenswrapper[4679]: I1202 10:23:04.095658 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:23:05 crc kubenswrapper[4679]: I1202 10:23:05.066299 4679 generic.go:334] "Generic (PLEG): container finished" podID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerID="69c6bfab086d5136a7ef6ce7b5f85730936be6d8960dd26d3436953a479e822c" exitCode=0 Dec 02 10:23:05 crc kubenswrapper[4679]: I1202 10:23:05.066351 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6jj7" event={"ID":"62c3d917-e199-4add-a42f-0cfdaba4dad4","Type":"ContainerDied","Data":"69c6bfab086d5136a7ef6ce7b5f85730936be6d8960dd26d3436953a479e822c"} Dec 02 10:23:05 crc kubenswrapper[4679]: I1202 10:23:05.068805 4679 generic.go:334] "Generic (PLEG): container finished" podID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerID="e4f0ba02e8f6356b3c1439b038f89ff5b5ebf9c5f9bc75ee9d132627626fd61f" exitCode=0 Dec 02 10:23:05 crc kubenswrapper[4679]: I1202 10:23:05.068837 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerDied","Data":"e4f0ba02e8f6356b3c1439b038f89ff5b5ebf9c5f9bc75ee9d132627626fd61f"} Dec 02 10:23:05 crc kubenswrapper[4679]: I1202 10:23:05.207839 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8w22n"] Dec 02 10:23:05 crc kubenswrapper[4679]: I1202 10:23:05.208221 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8w22n" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="registry-server" containerID="cri-o://a2cb47d9babbb438804391ffc71d92c6ee6812ef9865bb25a288e8c58a86dc08" gracePeriod=2 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.077510 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6jj7" event={"ID":"62c3d917-e199-4add-a42f-0cfdaba4dad4","Type":"ContainerDied","Data":"1861e2b8ab031f40b8d49ea73b32259858b4aeae4e70c4413bddde905d8ddaf8"} Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.077998 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1861e2b8ab031f40b8d49ea73b32259858b4aeae4e70c4413bddde905d8ddaf8" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.080776 4679 generic.go:334] "Generic (PLEG): container finished" podID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerID="a2cb47d9babbb438804391ffc71d92c6ee6812ef9865bb25a288e8c58a86dc08" exitCode=0 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.080829 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8w22n" event={"ID":"3bfee375-c85f-40ad-85fd-f16207687d6b","Type":"ContainerDied","Data":"a2cb47d9babbb438804391ffc71d92c6ee6812ef9865bb25a288e8c58a86dc08"} Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.109931 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.118682 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.245424 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-catalog-content\") pod \"62c3d917-e199-4add-a42f-0cfdaba4dad4\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.245516 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-utilities\") pod \"62c3d917-e199-4add-a42f-0cfdaba4dad4\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.245594 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-utilities\") pod \"ff5313a3-96e7-4e23-be49-8232b38d28c9\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.245628 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-catalog-content\") pod \"ff5313a3-96e7-4e23-be49-8232b38d28c9\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.245691 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4r5m\" (UniqueName: \"kubernetes.io/projected/62c3d917-e199-4add-a42f-0cfdaba4dad4-kube-api-access-j4r5m\") pod \"62c3d917-e199-4add-a42f-0cfdaba4dad4\" (UID: \"62c3d917-e199-4add-a42f-0cfdaba4dad4\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.245850 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph4xz\" (UniqueName: \"kubernetes.io/projected/ff5313a3-96e7-4e23-be49-8232b38d28c9-kube-api-access-ph4xz\") pod \"ff5313a3-96e7-4e23-be49-8232b38d28c9\" (UID: \"ff5313a3-96e7-4e23-be49-8232b38d28c9\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.247197 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-utilities" (OuterVolumeSpecName: "utilities") pod "ff5313a3-96e7-4e23-be49-8232b38d28c9" (UID: "ff5313a3-96e7-4e23-be49-8232b38d28c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.248629 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-utilities" (OuterVolumeSpecName: "utilities") pod "62c3d917-e199-4add-a42f-0cfdaba4dad4" (UID: "62c3d917-e199-4add-a42f-0cfdaba4dad4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.252472 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5313a3-96e7-4e23-be49-8232b38d28c9-kube-api-access-ph4xz" (OuterVolumeSpecName: "kube-api-access-ph4xz") pod "ff5313a3-96e7-4e23-be49-8232b38d28c9" (UID: "ff5313a3-96e7-4e23-be49-8232b38d28c9"). InnerVolumeSpecName "kube-api-access-ph4xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.253277 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c3d917-e199-4add-a42f-0cfdaba4dad4-kube-api-access-j4r5m" (OuterVolumeSpecName: "kube-api-access-j4r5m") pod "62c3d917-e199-4add-a42f-0cfdaba4dad4" (UID: "62c3d917-e199-4add-a42f-0cfdaba4dad4"). InnerVolumeSpecName "kube-api-access-j4r5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.296926 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff5313a3-96e7-4e23-be49-8232b38d28c9" (UID: "ff5313a3-96e7-4e23-be49-8232b38d28c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.301077 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62c3d917-e199-4add-a42f-0cfdaba4dad4" (UID: "62c3d917-e199-4add-a42f-0cfdaba4dad4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.347665 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph4xz\" (UniqueName: \"kubernetes.io/projected/ff5313a3-96e7-4e23-be49-8232b38d28c9-kube-api-access-ph4xz\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.347714 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.347732 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62c3d917-e199-4add-a42f-0cfdaba4dad4-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.347749 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.347762 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff5313a3-96e7-4e23-be49-8232b38d28c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.347777 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4r5m\" (UniqueName: \"kubernetes.io/projected/62c3d917-e199-4add-a42f-0cfdaba4dad4-kube-api-access-j4r5m\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.692148 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.852440 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjjld\" (UniqueName: \"kubernetes.io/projected/3bfee375-c85f-40ad-85fd-f16207687d6b-kube-api-access-hjjld\") pod \"3bfee375-c85f-40ad-85fd-f16207687d6b\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.852502 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-catalog-content\") pod \"3bfee375-c85f-40ad-85fd-f16207687d6b\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.852606 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-utilities\") pod \"3bfee375-c85f-40ad-85fd-f16207687d6b\" (UID: \"3bfee375-c85f-40ad-85fd-f16207687d6b\") " Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.853487 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-utilities" (OuterVolumeSpecName: "utilities") pod "3bfee375-c85f-40ad-85fd-f16207687d6b" (UID: "3bfee375-c85f-40ad-85fd-f16207687d6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.856405 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bfee375-c85f-40ad-85fd-f16207687d6b-kube-api-access-hjjld" (OuterVolumeSpecName: "kube-api-access-hjjld") pod "3bfee375-c85f-40ad-85fd-f16207687d6b" (UID: "3bfee375-c85f-40ad-85fd-f16207687d6b"). InnerVolumeSpecName "kube-api-access-hjjld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.880945 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bfee375-c85f-40ad-85fd-f16207687d6b" (UID: "3bfee375-c85f-40ad-85fd-f16207687d6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.953862 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjjld\" (UniqueName: \"kubernetes.io/projected/3bfee375-c85f-40ad-85fd-f16207687d6b-kube-api-access-hjjld\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.953900 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.953912 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfee375-c85f-40ad-85fd-f16207687d6b-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973271 4679 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973488 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="extract-content" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973499 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="extract-content" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973509 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="extract-content" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973515 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="extract-content" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973526 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="extract-content" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973532 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="extract-content" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973543 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="extract-utilities" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973550 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="extract-utilities" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973557 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973563 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973571 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973576 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973587 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="extract-utilities" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973592 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="extract-utilities" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973603 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="extract-utilities" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973608 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="extract-utilities" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973615 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd050d9-5ca0-4439-a758-ae85e28f334e" containerName="pruner" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973620 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd050d9-5ca0-4439-a758-ae85e28f334e" containerName="pruner" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973628 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614d686a-6f9a-47c2-93f0-0a6722e0d3fb" containerName="pruner" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973633 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="614d686a-6f9a-47c2-93f0-0a6722e0d3fb" containerName="pruner" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.973641 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973646 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973725 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fd050d9-5ca0-4439-a758-ae85e28f334e" containerName="pruner" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973737 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973744 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973752 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="614d686a-6f9a-47c2-93f0-0a6722e0d3fb" containerName="pruner" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.973762 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" containerName="registry-server" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974032 4679 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974242 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5" gracePeriod=15 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974277 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9" gracePeriod=15 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974275 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc" gracePeriod=15 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974343 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2" gracePeriod=15 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974346 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034" gracePeriod=15 Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.974446 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976393 4679 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.976616 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976632 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.976655 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976667 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.976682 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976694 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.976712 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976723 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.976747 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976760 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.976774 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976785 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976935 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976955 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976968 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.976986 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.977001 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.977013 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 10:23:06 crc kubenswrapper[4679]: E1202 10:23:06.977177 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 10:23:06 crc kubenswrapper[4679]: I1202 10:23:06.977192 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 10:23:07 crc kubenswrapper[4679]: E1202 10:23:07.009515 4679 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.088978 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8w22n" event={"ID":"3bfee375-c85f-40ad-85fd-f16207687d6b","Type":"ContainerDied","Data":"58f9950534d6ee3eee435bd7a25bb4fd66006ef6be19a910cf25c2dc1af15635"} Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.089024 4679 scope.go:117] "RemoveContainer" containerID="a2cb47d9babbb438804391ffc71d92c6ee6812ef9865bb25a288e8c58a86dc08" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.089159 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8w22n" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.090508 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.090738 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.093149 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dktz" event={"ID":"ff5313a3-96e7-4e23-be49-8232b38d28c9","Type":"ContainerDied","Data":"f6e43b6a587d0e6cd34d6d8ce420e7ce25babf61516fbc1bdf45c8869273c477"} Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.093207 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dktz" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.093208 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6jj7" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.093872 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.094113 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.094385 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.094588 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.094816 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.095011 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.095335 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.096114 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.140516 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.140958 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.141442 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.141903 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.142507 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.142862 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.143325 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.143573 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.147210 4679 scope.go:117] "RemoveContainer" containerID="0185b16bd01dcc02ae78ff581bd5ed8f87285d71372f47355ed678a8ee317f73" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156232 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156286 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156357 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156427 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156449 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156521 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156568 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.156592 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.165011 4679 scope.go:117] "RemoveContainer" containerID="ef0a1d61812d974c50e99536c2d5c30fe7e0c5d80be6abd4b9d26cd12fea3b35" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.179470 4679 scope.go:117] "RemoveContainer" containerID="e4f0ba02e8f6356b3c1439b038f89ff5b5ebf9c5f9bc75ee9d132627626fd61f" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.232960 4679 scope.go:117] "RemoveContainer" containerID="920e3d597e47adaff46a1622ce350d25d8faf577e7b3a19dc06287fda8412873" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.246325 4679 scope.go:117] "RemoveContainer" containerID="175a9296cd7cbd6bf473b1fe30e68741c5a32a61e28e7be29a69d9176d1d8334" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258041 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258084 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258103 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258140 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258165 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258180 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258206 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258221 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258279 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258336 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258360 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258380 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258399 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258417 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258436 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.258456 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: I1202 10:23:07.310704 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:07 crc kubenswrapper[4679]: W1202 10:23:07.326576 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-44d715205d4259f4e0f904aa2b04045402ab043b50db5edabbd08158fffef6f0 WatchSource:0}: Error finding container 44d715205d4259f4e0f904aa2b04045402ab043b50db5edabbd08158fffef6f0: Status 404 returned error can't find the container with id 44d715205d4259f4e0f904aa2b04045402ab043b50db5edabbd08158fffef6f0 Dec 02 10:23:07 crc kubenswrapper[4679]: E1202 10:23:07.329176 4679 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d5ee9d12a7c23 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 10:23:07.328715811 +0000 UTC m=+240.658854681,LastTimestamp:2025-12-02 10:23:07.328715811 +0000 UTC m=+240.658854681,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.099936 4679 generic.go:334] "Generic (PLEG): container finished" podID="a36424a7-1339-47b3-918a-4c51721b3476" containerID="4161ba6c8d8a4479c07c51fbdef4e842b214ffe4b87e349edd75a8ca16ac73a2" exitCode=0 Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.100009 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a36424a7-1339-47b3-918a-4c51721b3476","Type":"ContainerDied","Data":"4161ba6c8d8a4479c07c51fbdef4e842b214ffe4b87e349edd75a8ca16ac73a2"} Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.101371 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.101710 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.101924 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea"} Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.101982 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"44d715205d4259f4e0f904aa2b04045402ab043b50db5edabbd08158fffef6f0"} Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.101968 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.102431 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: E1202 10:23:08.102576 4679 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.102656 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.102913 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.103117 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.103389 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.103626 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.103815 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.103866 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.105215 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.105786 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc" exitCode=0 Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.105807 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034" exitCode=0 Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.105815 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9" exitCode=0 Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.105823 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2" exitCode=2 Dec 02 10:23:08 crc kubenswrapper[4679]: I1202 10:23:08.105871 4679 scope.go:117] "RemoveContainer" containerID="9c9f1b710d8586db0aad97e048e6700372344ef010f4a6bc5adf3e5a6ce6338f" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.117103 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 10:23:09 crc kubenswrapper[4679]: E1202 10:23:09.119047 4679 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.363370 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.364093 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.364249 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.364692 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.364983 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.365150 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.365287 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.365449 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.365653 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.365786 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.365919 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.366057 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.366191 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.486807 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36424a7-1339-47b3-918a-4c51721b3476-kube-api-access\") pod \"a36424a7-1339-47b3-918a-4c51721b3476\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.486868 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.486891 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.486908 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.486922 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-kubelet-dir\") pod \"a36424a7-1339-47b3-918a-4c51721b3476\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.486947 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-var-lock\") pod \"a36424a7-1339-47b3-918a-4c51721b3476\" (UID: \"a36424a7-1339-47b3-918a-4c51721b3476\") " Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.487205 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-var-lock" (OuterVolumeSpecName: "var-lock") pod "a36424a7-1339-47b3-918a-4c51721b3476" (UID: "a36424a7-1339-47b3-918a-4c51721b3476"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.487350 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.487379 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.487395 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a36424a7-1339-47b3-918a-4c51721b3476" (UID: "a36424a7-1339-47b3-918a-4c51721b3476"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.487393 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.493469 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36424a7-1339-47b3-918a-4c51721b3476-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a36424a7-1339-47b3-918a-4c51721b3476" (UID: "a36424a7-1339-47b3-918a-4c51721b3476"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.588974 4679 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.589036 4679 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.589056 4679 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a36424a7-1339-47b3-918a-4c51721b3476-var-lock\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.589074 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36424a7-1339-47b3-918a-4c51721b3476-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.589096 4679 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:09 crc kubenswrapper[4679]: I1202 10:23:09.589113 4679 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.128771 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a36424a7-1339-47b3-918a-4c51721b3476","Type":"ContainerDied","Data":"1d5a73b998338561c344b2cad9dc7ee130184029073dcc981821a5556dce29d2"} Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.128790 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.128818 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d5a73b998338561c344b2cad9dc7ee130184029073dcc981821a5556dce29d2" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.136970 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.137997 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5" exitCode=0 Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.138040 4679 scope.go:117] "RemoveContainer" containerID="7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.138166 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.143703 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.143991 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.144297 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.145176 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.145858 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.155108 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.155291 4679 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.155915 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.156433 4679 scope.go:117] "RemoveContainer" containerID="34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.157601 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.158025 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.173640 4679 scope.go:117] "RemoveContainer" containerID="68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.198939 4679 scope.go:117] "RemoveContainer" containerID="1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.211708 4679 scope.go:117] "RemoveContainer" containerID="8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.224507 4679 scope.go:117] "RemoveContainer" containerID="82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.251543 4679 scope.go:117] "RemoveContainer" containerID="7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc" Dec 02 10:23:10 crc kubenswrapper[4679]: E1202 10:23:10.252222 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\": container with ID starting with 7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc not found: ID does not exist" containerID="7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.252270 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc"} err="failed to get container status \"7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\": rpc error: code = NotFound desc = could not find container \"7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc\": container with ID starting with 7bf2251bd9fd491bef050f840b63a5fb76bff339008d57b46339737ccc0ef7bc not found: ID does not exist" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.252359 4679 scope.go:117] "RemoveContainer" containerID="34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034" Dec 02 10:23:10 crc kubenswrapper[4679]: E1202 10:23:10.252921 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\": container with ID starting with 34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034 not found: ID does not exist" containerID="34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.252955 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034"} err="failed to get container status \"34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\": rpc error: code = NotFound desc = could not find container \"34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034\": container with ID starting with 34ba9256da527e4cd0a8f9aea677a297aca50577ad5f57a1a0c0d8787553f034 not found: ID does not exist" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.252969 4679 scope.go:117] "RemoveContainer" containerID="68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9" Dec 02 10:23:10 crc kubenswrapper[4679]: E1202 10:23:10.253258 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\": container with ID starting with 68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9 not found: ID does not exist" containerID="68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.253293 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9"} err="failed to get container status \"68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\": rpc error: code = NotFound desc = could not find container \"68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9\": container with ID starting with 68e24974f8f594c68dd58237d739a0a6a2dc6ff353d13247b98413af2cae56f9 not found: ID does not exist" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.253330 4679 scope.go:117] "RemoveContainer" containerID="1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2" Dec 02 10:23:10 crc kubenswrapper[4679]: E1202 10:23:10.253565 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\": container with ID starting with 1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2 not found: ID does not exist" containerID="1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.253586 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2"} err="failed to get container status \"1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\": rpc error: code = NotFound desc = could not find container \"1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2\": container with ID starting with 1a5f1d28ce906dbb8e65614d1c5b4ae9ebdadaffcf23f0eaa5ca651165ce2ad2 not found: ID does not exist" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.253599 4679 scope.go:117] "RemoveContainer" containerID="8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5" Dec 02 10:23:10 crc kubenswrapper[4679]: E1202 10:23:10.254045 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\": container with ID starting with 8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5 not found: ID does not exist" containerID="8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.254073 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5"} err="failed to get container status \"8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\": rpc error: code = NotFound desc = could not find container \"8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5\": container with ID starting with 8f4fe40457fe91d619cd217ad13c7cbe05aa10850e293510a43ccb45c8ce0ac5 not found: ID does not exist" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.254091 4679 scope.go:117] "RemoveContainer" containerID="82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d" Dec 02 10:23:10 crc kubenswrapper[4679]: E1202 10:23:10.254472 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\": container with ID starting with 82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d not found: ID does not exist" containerID="82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.254504 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d"} err="failed to get container status \"82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\": rpc error: code = NotFound desc = could not find container \"82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d\": container with ID starting with 82346d9370ee2a14ed4cf526ceb7bb9aa44fa299255f39637a2e71fd453cf63d not found: ID does not exist" Dec 02 10:23:10 crc kubenswrapper[4679]: I1202 10:23:10.918175 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 02 10:23:11 crc kubenswrapper[4679]: E1202 10:23:11.416147 4679 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d5ee9d12a7c23 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 10:23:07.328715811 +0000 UTC m=+240.658854681,LastTimestamp:2025-12-02 10:23:07.328715811 +0000 UTC m=+240.658854681,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 10:23:12 crc kubenswrapper[4679]: E1202 10:23:12.828434 4679 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:12 crc kubenswrapper[4679]: E1202 10:23:12.829015 4679 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:12 crc kubenswrapper[4679]: E1202 10:23:12.829283 4679 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:12 crc kubenswrapper[4679]: E1202 10:23:12.829847 4679 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:12 crc kubenswrapper[4679]: E1202 10:23:12.830011 4679 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:12 crc kubenswrapper[4679]: I1202 10:23:12.830069 4679 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 02 10:23:12 crc kubenswrapper[4679]: E1202 10:23:12.830254 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="200ms" Dec 02 10:23:13 crc kubenswrapper[4679]: E1202 10:23:13.030842 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="400ms" Dec 02 10:23:13 crc kubenswrapper[4679]: E1202 10:23:13.432339 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="800ms" Dec 02 10:23:14 crc kubenswrapper[4679]: E1202 10:23:14.233851 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="1.6s" Dec 02 10:23:15 crc kubenswrapper[4679]: E1202 10:23:15.834435 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="3.2s" Dec 02 10:23:16 crc kubenswrapper[4679]: I1202 10:23:16.911254 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:16 crc kubenswrapper[4679]: I1202 10:23:16.911540 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:16 crc kubenswrapper[4679]: I1202 10:23:16.911841 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:16 crc kubenswrapper[4679]: I1202 10:23:16.912123 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.309959 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" containerName="oauth-openshift" containerID="cri-o://5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e" gracePeriod=15 Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.679632 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.680477 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.680786 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.681098 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.681327 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.681539 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791265 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-ocp-branding-template\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791343 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq5fb\" (UniqueName: \"kubernetes.io/projected/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-kube-api-access-dq5fb\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791371 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-login\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791390 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-router-certs\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791416 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-idp-0-file-data\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791446 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-trusted-ca-bundle\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791479 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-dir\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792360 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-policies\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792400 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-service-ca\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.791635 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792268 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792445 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-serving-cert\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792512 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-error\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792558 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-cliconfig\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792592 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-session\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792627 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-provider-selection\") pod \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\" (UID: \"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76\") " Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792887 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792914 4679 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.792913 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.793041 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.793216 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.797319 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.797549 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.797987 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.798088 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-kube-api-access-dq5fb" (OuterVolumeSpecName: "kube-api-access-dq5fb") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "kube-api-access-dq5fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.798128 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.798320 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.799080 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.799861 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.802026 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" (UID: "975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894526 4679 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894590 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894616 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894637 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894657 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894672 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894688 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894704 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894715 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq5fb\" (UniqueName: \"kubernetes.io/projected/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-kube-api-access-dq5fb\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894731 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894746 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:17 crc kubenswrapper[4679]: I1202 10:23:17.894756 4679 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.185826 4679 generic.go:334] "Generic (PLEG): container finished" podID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" containerID="5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e" exitCode=0 Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.185872 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" event={"ID":"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76","Type":"ContainerDied","Data":"5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e"} Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.185901 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" event={"ID":"975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76","Type":"ContainerDied","Data":"ed6a729ba2d992773cd58da28e07d48ccb1ab06a1e1db492907f27af8d2d794d"} Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.185917 4679 scope.go:117] "RemoveContainer" containerID="5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.186029 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.186943 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.187545 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.187749 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.188497 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.188883 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.200149 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.200472 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.200817 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.201265 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.201521 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.206780 4679 scope.go:117] "RemoveContainer" containerID="5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e" Dec 02 10:23:18 crc kubenswrapper[4679]: E1202 10:23:18.207259 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e\": container with ID starting with 5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e not found: ID does not exist" containerID="5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e" Dec 02 10:23:18 crc kubenswrapper[4679]: I1202 10:23:18.207334 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e"} err="failed to get container status \"5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e\": rpc error: code = NotFound desc = could not find container \"5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e\": container with ID starting with 5ffea4da6335cc048db7364dfda03da29586632777116eb510ceb4e7b67f445e not found: ID does not exist" Dec 02 10:23:19 crc kubenswrapper[4679]: E1202 10:23:19.036496 4679 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="6.4s" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.204860 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.204908 4679 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514" exitCode=1 Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.204936 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514"} Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.205436 4679 scope.go:117] "RemoveContainer" containerID="82d9dbecfda17b743af800639ebcc0430ac0af04cc6a98cae4c40c7b45b49514" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.205862 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.206203 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.206411 4679 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.206584 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.206866 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:20 crc kubenswrapper[4679]: I1202 10:23:20.207202 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.212094 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.212497 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c1f2e27455e4eea6860fd02a0666c080afb82882582ac431c88784660b4e4bb1"} Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.213391 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.213762 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.214098 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.214395 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.214622 4679 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.214841 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: E1202 10:23:21.417533 4679 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d5ee9d12a7c23 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 10:23:07.328715811 +0000 UTC m=+240.658854681,LastTimestamp:2025-12-02 10:23:07.328715811 +0000 UTC m=+240.658854681,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.907736 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.908789 4679 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.909581 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.910152 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.910460 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.910784 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.911120 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.921499 4679 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.921535 4679 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:21 crc kubenswrapper[4679]: E1202 10:23:21.921949 4679 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:21 crc kubenswrapper[4679]: I1202 10:23:21.922476 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:21 crc kubenswrapper[4679]: W1202 10:23:21.941521 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-62a26f3bc22113b14bbefeae13102c46b8ccea9e46fba3de8b58dda5a1c4d99c WatchSource:0}: Error finding container 62a26f3bc22113b14bbefeae13102c46b8ccea9e46fba3de8b58dda5a1c4d99c: Status 404 returned error can't find the container with id 62a26f3bc22113b14bbefeae13102c46b8ccea9e46fba3de8b58dda5a1c4d99c Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.222091 4679 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="cab07319a4fb1d090ee0057a7db88d84761ce1cb4ee40d3be61216a5cc28ce96" exitCode=0 Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.222150 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"cab07319a4fb1d090ee0057a7db88d84761ce1cb4ee40d3be61216a5cc28ce96"} Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.222193 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"62a26f3bc22113b14bbefeae13102c46b8ccea9e46fba3de8b58dda5a1c4d99c"} Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.222577 4679 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.222596 4679 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.223273 4679 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:22 crc kubenswrapper[4679]: E1202 10:23:22.223359 4679 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.223675 4679 status_manager.go:851] "Failed to get status for pod" podUID="a36424a7-1339-47b3-918a-4c51721b3476" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.223940 4679 status_manager.go:851] "Failed to get status for pod" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" pod="openshift-marketplace/certified-operators-h6jj7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-h6jj7\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.224239 4679 status_manager.go:851] "Failed to get status for pod" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" pod="openshift-marketplace/redhat-marketplace-8w22n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-8w22n\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.224489 4679 status_manager.go:851] "Failed to get status for pod" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" pod="openshift-marketplace/community-operators-4dktz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-4dktz\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:22 crc kubenswrapper[4679]: I1202 10:23:22.224786 4679 status_manager.go:851] "Failed to get status for pod" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" pod="openshift-authentication/oauth-openshift-558db77b4-68gpm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-68gpm\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 02 10:23:23 crc kubenswrapper[4679]: I1202 10:23:23.232780 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"75c9cf6d3ab905d38b2f730b3881d62f6bad9f9402a73b29c576c6ca782886da"} Dec 02 10:23:23 crc kubenswrapper[4679]: I1202 10:23:23.233066 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"998b2b73646ae40aeb7ad35cc97568e1371db9b793352f78c3d4267fda890530"} Dec 02 10:23:23 crc kubenswrapper[4679]: I1202 10:23:23.233082 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"81e134ae71ea32ff92e8dd6fb66f6c2215741b941d182d433679500bb3257d76"} Dec 02 10:23:23 crc kubenswrapper[4679]: I1202 10:23:23.233094 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0adcfa18a1b7f2b0b7e5860d312a6b0236b76cae426636909a31bdd092b92529"} Dec 02 10:23:23 crc kubenswrapper[4679]: I1202 10:23:23.672720 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:23:24 crc kubenswrapper[4679]: I1202 10:23:24.170553 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:23:24 crc kubenswrapper[4679]: I1202 10:23:24.174431 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:23:24 crc kubenswrapper[4679]: I1202 10:23:24.239196 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0f3047631b1923f39f22f9b76c5fe27b0211fa6d918887f05628b50c35f6c91d"} Dec 02 10:23:24 crc kubenswrapper[4679]: I1202 10:23:24.239534 4679 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:24 crc kubenswrapper[4679]: I1202 10:23:24.239552 4679 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:26 crc kubenswrapper[4679]: I1202 10:23:26.923095 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:26 crc kubenswrapper[4679]: I1202 10:23:26.923421 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:26 crc kubenswrapper[4679]: I1202 10:23:26.929183 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:29 crc kubenswrapper[4679]: I1202 10:23:29.250376 4679 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:29 crc kubenswrapper[4679]: I1202 10:23:29.300810 4679 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7b4bc7ea-3499-452f-9c11-d2326ef43ecd" Dec 02 10:23:30 crc kubenswrapper[4679]: I1202 10:23:30.266110 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:30 crc kubenswrapper[4679]: I1202 10:23:30.266218 4679 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:30 crc kubenswrapper[4679]: I1202 10:23:30.266250 4679 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:30 crc kubenswrapper[4679]: I1202 10:23:30.270040 4679 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7b4bc7ea-3499-452f-9c11-d2326ef43ecd" Dec 02 10:23:31 crc kubenswrapper[4679]: I1202 10:23:31.273780 4679 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:31 crc kubenswrapper[4679]: I1202 10:23:31.273822 4679 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:31 crc kubenswrapper[4679]: I1202 10:23:31.276774 4679 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7b4bc7ea-3499-452f-9c11-d2326ef43ecd" Dec 02 10:23:33 crc kubenswrapper[4679]: I1202 10:23:33.677285 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 10:23:38 crc kubenswrapper[4679]: I1202 10:23:38.986184 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 02 10:23:38 crc kubenswrapper[4679]: I1202 10:23:38.986964 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.062739 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.075534 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.138867 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.311048 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.406724 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.422015 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.625274 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.719616 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 02 10:23:39 crc kubenswrapper[4679]: I1202 10:23:39.764569 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.043015 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.053669 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.314549 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.544068 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.649944 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.800182 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.883774 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 02 10:23:40 crc kubenswrapper[4679]: I1202 10:23:40.997052 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.068943 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.117948 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.508818 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.523003 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.585810 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.688160 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.911093 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.925997 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.953872 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 02 10:23:41 crc kubenswrapper[4679]: I1202 10:23:41.970687 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.193102 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.220690 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.243669 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.315605 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.349405 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.405551 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.417541 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.437866 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.683209 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.704552 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.772958 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.801778 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 02 10:23:42 crc kubenswrapper[4679]: I1202 10:23:42.893114 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.009042 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.071082 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.111949 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.171996 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.252958 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.264251 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.333165 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.364010 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.413612 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.450188 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.650700 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.709738 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.715974 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.716187 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.735591 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.759374 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 02 10:23:43 crc kubenswrapper[4679]: I1202 10:23:43.829484 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.059115 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.062176 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.081050 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.175654 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.214736 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.319810 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.397788 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.477412 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.491197 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.585205 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.585435 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.588503 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.594654 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.692153 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.781747 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.802387 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.908794 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 02 10:23:44 crc kubenswrapper[4679]: I1202 10:23:44.963616 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.038429 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.163659 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.209168 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.245231 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.250851 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.262903 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.299835 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.329029 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.407704 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.487917 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.655762 4679 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.719919 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.774103 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.828296 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 02 10:23:45 crc kubenswrapper[4679]: I1202 10:23:45.893957 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.025486 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.042885 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.073413 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.074054 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.079060 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.104653 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.168297 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.194576 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.225109 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.241473 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.290158 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.437416 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.470128 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.507676 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.525585 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.712818 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.737591 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.738077 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.779751 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.783150 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.852293 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.900848 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.927028 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 02 10:23:46 crc kubenswrapper[4679]: I1202 10:23:46.969647 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.035092 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.035100 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.036210 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.126431 4679 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.200212 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.289991 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.322182 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.410475 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.411985 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.428110 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.501189 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.571463 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.728587 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.789688 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.886940 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.893097 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.928479 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 02 10:23:47 crc kubenswrapper[4679]: I1202 10:23:47.929865 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.005802 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.094736 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.165841 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.180844 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.238617 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.294240 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.317743 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.333290 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.403997 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.417364 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.519372 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.519838 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.679490 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.696615 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.717848 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.783394 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.846255 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.910727 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 02 10:23:48 crc kubenswrapper[4679]: I1202 10:23:48.989485 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.183114 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.200192 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.215481 4679 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.368144 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.443492 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.553289 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.578977 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.653914 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.668200 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.755119 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.801482 4679 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.833533 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.930509 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 02 10:23:49 crc kubenswrapper[4679]: I1202 10:23:49.949466 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.007904 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.032122 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.099334 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.108686 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.120833 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.140494 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.186718 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.481681 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.502835 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.510668 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.619840 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.650786 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 02 10:23:50 crc kubenswrapper[4679]: I1202 10:23:50.661410 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.012033 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.075208 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.100862 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.109907 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.125849 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.212922 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.258480 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.258753 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.375408 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.385547 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.407984 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.410658 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.493629 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.559831 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.579963 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.625603 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.658175 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.813957 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.837377 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.864764 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.892330 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.903736 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.908282 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.931219 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.932156 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 10:23:51 crc kubenswrapper[4679]: I1202 10:23:51.974380 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.012868 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.117925 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.125903 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.155911 4679 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.181760 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.290589 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.358632 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.487570 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.536218 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.693425 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.733183 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.774395 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 02 10:23:52 crc kubenswrapper[4679]: I1202 10:23:52.863374 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.028665 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.085851 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.356776 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.704171 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.710221 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.795956 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.812617 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.912434 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 02 10:23:53 crc kubenswrapper[4679]: I1202 10:23:53.997287 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.213824 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.439856 4679 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.444817 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4dktz","openshift-authentication/oauth-openshift-558db77b4-68gpm","openshift-marketplace/redhat-marketplace-8w22n","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/certified-operators-h6jj7"] Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.444893 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6f67d677dd-hdg87","openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 10:23:54 crc kubenswrapper[4679]: E1202 10:23:54.445097 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" containerName="oauth-openshift" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.445114 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" containerName="oauth-openshift" Dec 02 10:23:54 crc kubenswrapper[4679]: E1202 10:23:54.445130 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36424a7-1339-47b3-918a-4c51721b3476" containerName="installer" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.445138 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36424a7-1339-47b3-918a-4c51721b3476" containerName="installer" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.445253 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" containerName="oauth-openshift" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.445269 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36424a7-1339-47b3-918a-4c51721b3476" containerName="installer" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.447795 4679 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.447825 4679 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="93a1d8d7-802a-4246-84b9-7cb1880de962" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.448327 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.450240 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.451195 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.452518 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.452552 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.452738 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.452972 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.452825 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.453116 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.453156 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.454298 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.454395 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.454750 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.455725 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.462635 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.469156 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.481668 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.482195 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=25.482171636 podStartE2EDuration="25.482171636s" podCreationTimestamp="2025-12-02 10:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:23:54.476533071 +0000 UTC m=+287.806671931" watchObservedRunningTime="2025-12-02 10:23:54.482171636 +0000 UTC m=+287.812310496" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527540 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527589 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527615 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527633 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527657 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527851 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g76t\" (UniqueName: \"kubernetes.io/projected/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-kube-api-access-7g76t\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527934 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-session\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.527977 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.528077 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-login\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.528134 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.528219 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-audit-policies\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.528266 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.528290 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-audit-dir\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.528329 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-error\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629269 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629337 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-audit-policies\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629361 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629381 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-audit-dir\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629413 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-error\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629451 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629478 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629507 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629533 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629571 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629599 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g76t\" (UniqueName: \"kubernetes.io/projected/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-kube-api-access-7g76t\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629628 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-session\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629650 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629686 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-login\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.629505 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-audit-dir\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.630624 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.630746 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-audit-policies\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.631287 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.631923 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.636862 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.636879 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-session\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.636945 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.637034 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.637092 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-login\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.637568 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.645692 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.646556 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-v4-0-config-user-template-error\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.646736 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g76t\" (UniqueName: \"kubernetes.io/projected/99a5f5a7-3ac3-46f0-88ab-65ba5da686cd-kube-api-access-7g76t\") pod \"oauth-openshift-6f67d677dd-hdg87\" (UID: \"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd\") " pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.667415 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.694346 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.700997 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.771965 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.786982 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.915615 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bfee375-c85f-40ad-85fd-f16207687d6b" path="/var/lib/kubelet/pods/3bfee375-c85f-40ad-85fd-f16207687d6b/volumes" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.916341 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c3d917-e199-4add-a42f-0cfdaba4dad4" path="/var/lib/kubelet/pods/62c3d917-e199-4add-a42f-0cfdaba4dad4/volumes" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.917056 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76" path="/var/lib/kubelet/pods/975c5cdb-bf9b-4598-9b66-7b4d1e6ebc76/volumes" Dec 02 10:23:54 crc kubenswrapper[4679]: I1202 10:23:54.918154 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5313a3-96e7-4e23-be49-8232b38d28c9" path="/var/lib/kubelet/pods/ff5313a3-96e7-4e23-be49-8232b38d28c9/volumes" Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.162697 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.180975 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6f67d677dd-hdg87"] Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.397443 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" event={"ID":"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd","Type":"ContainerStarted","Data":"5ec29a4a76311c866022a092b8066cc5959e79509aa5b40a2f744c297e35f9aa"} Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.397835 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" event={"ID":"99a5f5a7-3ac3-46f0-88ab-65ba5da686cd","Type":"ContainerStarted","Data":"2145b034123dc8be3234c20b7e77680bf47b4cfa9a8276c0ce547305d21c9eae"} Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.398608 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.400065 4679 patch_prober.go:28] interesting pod/oauth-openshift-6f67d677dd-hdg87 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" start-of-body= Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.400112 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" podUID="99a5f5a7-3ac3-46f0-88ab-65ba5da686cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": dial tcp 10.217.0.56:6443: connect: connection refused" Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.403711 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.420510 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" podStartSLOduration=63.420472524 podStartE2EDuration="1m3.420472524s" podCreationTimestamp="2025-12-02 10:22:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:23:55.418901368 +0000 UTC m=+288.749040228" watchObservedRunningTime="2025-12-02 10:23:55.420472524 +0000 UTC m=+288.750611384" Dec 02 10:23:55 crc kubenswrapper[4679]: I1202 10:23:55.864782 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 02 10:23:56 crc kubenswrapper[4679]: I1202 10:23:56.407697 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6f67d677dd-hdg87" Dec 02 10:24:02 crc kubenswrapper[4679]: I1202 10:24:02.845138 4679 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 10:24:02 crc kubenswrapper[4679]: I1202 10:24:02.846035 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea" gracePeriod=5 Dec 02 10:24:06 crc kubenswrapper[4679]: I1202 10:24:06.790918 4679 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.416944 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.417402 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.494871 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.494938 4679 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea" exitCode=137 Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.494987 4679 scope.go:117] "RemoveContainer" containerID="2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.495046 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499559 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499614 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499639 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499662 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499661 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499703 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499705 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499722 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.499743 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.500039 4679 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.500056 4679 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.500066 4679 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.500077 4679 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.507026 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.510912 4679 scope.go:117] "RemoveContainer" containerID="2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea" Dec 02 10:24:08 crc kubenswrapper[4679]: E1202 10:24:08.512679 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea\": container with ID starting with 2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea not found: ID does not exist" containerID="2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.512719 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea"} err="failed to get container status \"2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea\": rpc error: code = NotFound desc = could not find container \"2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea\": container with ID starting with 2037967d8535778b41310a906f821f0ddfd10f73edff09428ceabebb88221aea not found: ID does not exist" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.601514 4679 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:08 crc kubenswrapper[4679]: I1202 10:24:08.916365 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 02 10:24:12 crc kubenswrapper[4679]: I1202 10:24:12.523920 4679 generic.go:334] "Generic (PLEG): container finished" podID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerID="f09cea1eb410b7707a4fe11c8b8df1052a5fd0bfabcc95d71f61365de9beae3a" exitCode=0 Dec 02 10:24:12 crc kubenswrapper[4679]: I1202 10:24:12.523992 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" event={"ID":"a180dc8a-e1e6-4da9-bf62-e973ef9d4611","Type":"ContainerDied","Data":"f09cea1eb410b7707a4fe11c8b8df1052a5fd0bfabcc95d71f61365de9beae3a"} Dec 02 10:24:12 crc kubenswrapper[4679]: I1202 10:24:12.524928 4679 scope.go:117] "RemoveContainer" containerID="f09cea1eb410b7707a4fe11c8b8df1052a5fd0bfabcc95d71f61365de9beae3a" Dec 02 10:24:13 crc kubenswrapper[4679]: I1202 10:24:13.533324 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" event={"ID":"a180dc8a-e1e6-4da9-bf62-e973ef9d4611","Type":"ContainerStarted","Data":"5fab1dc017b001583cc28843c17075f09fb8940ef13a6c8743f503d8b05274ba"} Dec 02 10:24:13 crc kubenswrapper[4679]: I1202 10:24:13.533695 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:24:13 crc kubenswrapper[4679]: I1202 10:24:13.537427 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:24:27 crc kubenswrapper[4679]: I1202 10:24:27.984770 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 02 10:24:36 crc kubenswrapper[4679]: I1202 10:24:36.245699 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jr6tk"] Dec 02 10:24:36 crc kubenswrapper[4679]: I1202 10:24:36.246172 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jr6tk" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="registry-server" containerID="cri-o://46dbd29969ab03b89fbd2b6a4d4cbb3d5d826f0a2c441428af64cb8044aedca5" gracePeriod=2 Dec 02 10:24:36 crc kubenswrapper[4679]: I1202 10:24:36.653029 4679 generic.go:334] "Generic (PLEG): container finished" podID="351b1389-f00d-4434-b58e-34a76ab674d6" containerID="46dbd29969ab03b89fbd2b6a4d4cbb3d5d826f0a2c441428af64cb8044aedca5" exitCode=0 Dec 02 10:24:36 crc kubenswrapper[4679]: I1202 10:24:36.653088 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerDied","Data":"46dbd29969ab03b89fbd2b6a4d4cbb3d5d826f0a2c441428af64cb8044aedca5"} Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.059326 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.152798 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbf6j\" (UniqueName: \"kubernetes.io/projected/351b1389-f00d-4434-b58e-34a76ab674d6-kube-api-access-mbf6j\") pod \"351b1389-f00d-4434-b58e-34a76ab674d6\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.152895 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-catalog-content\") pod \"351b1389-f00d-4434-b58e-34a76ab674d6\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.152966 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-utilities\") pod \"351b1389-f00d-4434-b58e-34a76ab674d6\" (UID: \"351b1389-f00d-4434-b58e-34a76ab674d6\") " Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.153922 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-utilities" (OuterVolumeSpecName: "utilities") pod "351b1389-f00d-4434-b58e-34a76ab674d6" (UID: "351b1389-f00d-4434-b58e-34a76ab674d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.158257 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351b1389-f00d-4434-b58e-34a76ab674d6-kube-api-access-mbf6j" (OuterVolumeSpecName: "kube-api-access-mbf6j") pod "351b1389-f00d-4434-b58e-34a76ab674d6" (UID: "351b1389-f00d-4434-b58e-34a76ab674d6"). InnerVolumeSpecName "kube-api-access-mbf6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.255095 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbf6j\" (UniqueName: \"kubernetes.io/projected/351b1389-f00d-4434-b58e-34a76ab674d6-kube-api-access-mbf6j\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.255133 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.287271 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "351b1389-f00d-4434-b58e-34a76ab674d6" (UID: "351b1389-f00d-4434-b58e-34a76ab674d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.356451 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351b1389-f00d-4434-b58e-34a76ab674d6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.662200 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr6tk" event={"ID":"351b1389-f00d-4434-b58e-34a76ab674d6","Type":"ContainerDied","Data":"ad3aa2f18dc11344862e2deb61019573fec37fbcc9d418af585c1d1bb2de7614"} Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.662245 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jr6tk" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.662289 4679 scope.go:117] "RemoveContainer" containerID="46dbd29969ab03b89fbd2b6a4d4cbb3d5d826f0a2c441428af64cb8044aedca5" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.680748 4679 scope.go:117] "RemoveContainer" containerID="2936e1212b4ace676c019115c2bcbe9a0e5022c52da16dcd6ea2e1a0e1b21832" Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.691121 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jr6tk"] Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.693935 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jr6tk"] Dec 02 10:24:37 crc kubenswrapper[4679]: I1202 10:24:37.711464 4679 scope.go:117] "RemoveContainer" containerID="3b6634a93fb85a66ed330f0e60b05da30f42b3910ad4872424f5a6f9fc0d716d" Dec 02 10:24:38 crc kubenswrapper[4679]: I1202 10:24:38.915857 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" path="/var/lib/kubelet/pods/351b1389-f00d-4434-b58e-34a76ab674d6/volumes" Dec 02 10:24:50 crc kubenswrapper[4679]: I1202 10:24:50.804595 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cx5cx"] Dec 02 10:24:50 crc kubenswrapper[4679]: I1202 10:24:50.805383 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" podUID="9821b1bc-27fb-4720-afce-16072b779105" containerName="controller-manager" containerID="cri-o://3381b86a8f7f228e39088ef40f6975e0504d2e4766c9f34cc54aeabbad8ce9d6" gracePeriod=30 Dec 02 10:24:50 crc kubenswrapper[4679]: I1202 10:24:50.916279 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd"] Dec 02 10:24:50 crc kubenswrapper[4679]: I1202 10:24:50.916515 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" podUID="7caf2887-b32b-4d3d-9c93-a116076dde2c" containerName="route-controller-manager" containerID="cri-o://6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3" gracePeriod=30 Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.240766 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.330734 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/7caf2887-b32b-4d3d-9c93-a116076dde2c-kube-api-access-9jzfh\") pod \"7caf2887-b32b-4d3d-9c93-a116076dde2c\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.330838 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-client-ca\") pod \"7caf2887-b32b-4d3d-9c93-a116076dde2c\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.330866 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-config\") pod \"7caf2887-b32b-4d3d-9c93-a116076dde2c\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.330908 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7caf2887-b32b-4d3d-9c93-a116076dde2c-serving-cert\") pod \"7caf2887-b32b-4d3d-9c93-a116076dde2c\" (UID: \"7caf2887-b32b-4d3d-9c93-a116076dde2c\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.331708 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-client-ca" (OuterVolumeSpecName: "client-ca") pod "7caf2887-b32b-4d3d-9c93-a116076dde2c" (UID: "7caf2887-b32b-4d3d-9c93-a116076dde2c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.331859 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-config" (OuterVolumeSpecName: "config") pod "7caf2887-b32b-4d3d-9c93-a116076dde2c" (UID: "7caf2887-b32b-4d3d-9c93-a116076dde2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.339484 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7caf2887-b32b-4d3d-9c93-a116076dde2c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7caf2887-b32b-4d3d-9c93-a116076dde2c" (UID: "7caf2887-b32b-4d3d-9c93-a116076dde2c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.339834 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7caf2887-b32b-4d3d-9c93-a116076dde2c-kube-api-access-9jzfh" (OuterVolumeSpecName: "kube-api-access-9jzfh") pod "7caf2887-b32b-4d3d-9c93-a116076dde2c" (UID: "7caf2887-b32b-4d3d-9c93-a116076dde2c"). InnerVolumeSpecName "kube-api-access-9jzfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.433163 4679 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.433221 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7caf2887-b32b-4d3d-9c93-a116076dde2c-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.433240 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7caf2887-b32b-4d3d-9c93-a116076dde2c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.433261 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/7caf2887-b32b-4d3d-9c93-a116076dde2c-kube-api-access-9jzfh\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.738406 4679 generic.go:334] "Generic (PLEG): container finished" podID="7caf2887-b32b-4d3d-9c93-a116076dde2c" containerID="6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3" exitCode=0 Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.738539 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" event={"ID":"7caf2887-b32b-4d3d-9c93-a116076dde2c","Type":"ContainerDied","Data":"6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3"} Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.738599 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" event={"ID":"7caf2887-b32b-4d3d-9c93-a116076dde2c","Type":"ContainerDied","Data":"ba18f9ff04c3e78332adcbe93af0667c17e6f0b70d95597e341599be4072bce0"} Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.738623 4679 scope.go:117] "RemoveContainer" containerID="6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.738635 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.744341 4679 generic.go:334] "Generic (PLEG): container finished" podID="9821b1bc-27fb-4720-afce-16072b779105" containerID="3381b86a8f7f228e39088ef40f6975e0504d2e4766c9f34cc54aeabbad8ce9d6" exitCode=0 Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.744391 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" event={"ID":"9821b1bc-27fb-4720-afce-16072b779105","Type":"ContainerDied","Data":"3381b86a8f7f228e39088ef40f6975e0504d2e4766c9f34cc54aeabbad8ce9d6"} Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.765634 4679 scope.go:117] "RemoveContainer" containerID="6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3" Dec 02 10:24:51 crc kubenswrapper[4679]: E1202 10:24:51.766391 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3\": container with ID starting with 6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3 not found: ID does not exist" containerID="6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.766433 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3"} err="failed to get container status \"6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3\": rpc error: code = NotFound desc = could not find container \"6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3\": container with ID starting with 6402cac3775390e691097782f91016c1f5ccdb2934c6ffb616f59c397b1e1fb3 not found: ID does not exist" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.786087 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd"] Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.790662 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p49dd"] Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.861504 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.985958 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-proxy-ca-bundles\") pod \"9821b1bc-27fb-4720-afce-16072b779105\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.990635 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5mcz\" (UniqueName: \"kubernetes.io/projected/9821b1bc-27fb-4720-afce-16072b779105-kube-api-access-d5mcz\") pod \"9821b1bc-27fb-4720-afce-16072b779105\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.990693 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9821b1bc-27fb-4720-afce-16072b779105-serving-cert\") pod \"9821b1bc-27fb-4720-afce-16072b779105\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.990729 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-config\") pod \"9821b1bc-27fb-4720-afce-16072b779105\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.990766 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-client-ca\") pod \"9821b1bc-27fb-4720-afce-16072b779105\" (UID: \"9821b1bc-27fb-4720-afce-16072b779105\") " Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.987975 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9821b1bc-27fb-4720-afce-16072b779105" (UID: "9821b1bc-27fb-4720-afce-16072b779105"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.991393 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-client-ca" (OuterVolumeSpecName: "client-ca") pod "9821b1bc-27fb-4720-afce-16072b779105" (UID: "9821b1bc-27fb-4720-afce-16072b779105"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.995047 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-config" (OuterVolumeSpecName: "config") pod "9821b1bc-27fb-4720-afce-16072b779105" (UID: "9821b1bc-27fb-4720-afce-16072b779105"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:24:51 crc kubenswrapper[4679]: I1202 10:24:51.995072 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9821b1bc-27fb-4720-afce-16072b779105-kube-api-access-d5mcz" (OuterVolumeSpecName: "kube-api-access-d5mcz") pod "9821b1bc-27fb-4720-afce-16072b779105" (UID: "9821b1bc-27fb-4720-afce-16072b779105"). InnerVolumeSpecName "kube-api-access-d5mcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.000430 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9821b1bc-27fb-4720-afce-16072b779105-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9821b1bc-27fb-4720-afce-16072b779105" (UID: "9821b1bc-27fb-4720-afce-16072b779105"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.092412 4679 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.092464 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5mcz\" (UniqueName: \"kubernetes.io/projected/9821b1bc-27fb-4720-afce-16072b779105-kube-api-access-d5mcz\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.092476 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9821b1bc-27fb-4720-afce-16072b779105-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.092484 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.092494 4679 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9821b1bc-27fb-4720-afce-16072b779105-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.430619 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-f9rdk"] Dec 02 10:24:52 crc kubenswrapper[4679]: E1202 10:24:52.431238 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7caf2887-b32b-4d3d-9c93-a116076dde2c" containerName="route-controller-manager" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431264 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7caf2887-b32b-4d3d-9c93-a116076dde2c" containerName="route-controller-manager" Dec 02 10:24:52 crc kubenswrapper[4679]: E1202 10:24:52.431295 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="extract-utilities" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431351 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="extract-utilities" Dec 02 10:24:52 crc kubenswrapper[4679]: E1202 10:24:52.431377 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="extract-content" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431393 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="extract-content" Dec 02 10:24:52 crc kubenswrapper[4679]: E1202 10:24:52.431428 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="registry-server" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431443 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="registry-server" Dec 02 10:24:52 crc kubenswrapper[4679]: E1202 10:24:52.431463 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431480 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 10:24:52 crc kubenswrapper[4679]: E1202 10:24:52.431504 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9821b1bc-27fb-4720-afce-16072b779105" containerName="controller-manager" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431517 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="9821b1bc-27fb-4720-afce-16072b779105" containerName="controller-manager" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431693 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="351b1389-f00d-4434-b58e-34a76ab674d6" containerName="registry-server" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431710 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431734 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7caf2887-b32b-4d3d-9c93-a116076dde2c" containerName="route-controller-manager" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.431754 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="9821b1bc-27fb-4720-afce-16072b779105" containerName="controller-manager" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.432368 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.436987 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-f9rdk"] Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.497757 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwxw7\" (UniqueName: \"kubernetes.io/projected/5061ae44-8130-4aaf-8258-1bed986489af-kube-api-access-wwxw7\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.497836 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5061ae44-8130-4aaf-8258-1bed986489af-serving-cert\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.497876 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-proxy-ca-bundles\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.497899 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-client-ca\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.497949 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-config\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.598880 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwxw7\" (UniqueName: \"kubernetes.io/projected/5061ae44-8130-4aaf-8258-1bed986489af-kube-api-access-wwxw7\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.598932 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5061ae44-8130-4aaf-8258-1bed986489af-serving-cert\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.598955 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-proxy-ca-bundles\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.598971 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-client-ca\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.599013 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-config\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.600296 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-client-ca\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.600396 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-config\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.601258 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-proxy-ca-bundles\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.603679 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5061ae44-8130-4aaf-8258-1bed986489af-serving-cert\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.625969 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwxw7\" (UniqueName: \"kubernetes.io/projected/5061ae44-8130-4aaf-8258-1bed986489af-kube-api-access-wwxw7\") pod \"controller-manager-658fd5994d-f9rdk\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.642336 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q"] Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.642984 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.647050 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.647920 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.648065 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.648111 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.648254 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.648382 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.658917 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q"] Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.700246 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3637b659-50e8-4bd7-87a5-918a50b0e73e-client-ca\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.700324 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3637b659-50e8-4bd7-87a5-918a50b0e73e-serving-cert\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.700452 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg2w7\" (UniqueName: \"kubernetes.io/projected/3637b659-50e8-4bd7-87a5-918a50b0e73e-kube-api-access-hg2w7\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.700495 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3637b659-50e8-4bd7-87a5-918a50b0e73e-config\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.752216 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.752739 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" event={"ID":"9821b1bc-27fb-4720-afce-16072b779105","Type":"ContainerDied","Data":"356febb2b813d8eee1812ae7ab6b8b90d9336a3cf0dda7aaf0d49085ddd1f085"} Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.752847 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cx5cx" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.752856 4679 scope.go:117] "RemoveContainer" containerID="3381b86a8f7f228e39088ef40f6975e0504d2e4766c9f34cc54aeabbad8ce9d6" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.802820 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3637b659-50e8-4bd7-87a5-918a50b0e73e-serving-cert\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.802909 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg2w7\" (UniqueName: \"kubernetes.io/projected/3637b659-50e8-4bd7-87a5-918a50b0e73e-kube-api-access-hg2w7\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.802954 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3637b659-50e8-4bd7-87a5-918a50b0e73e-config\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.803070 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3637b659-50e8-4bd7-87a5-918a50b0e73e-client-ca\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.804438 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3637b659-50e8-4bd7-87a5-918a50b0e73e-client-ca\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.806398 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3637b659-50e8-4bd7-87a5-918a50b0e73e-config\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.808831 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cx5cx"] Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.812943 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3637b659-50e8-4bd7-87a5-918a50b0e73e-serving-cert\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.813773 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cx5cx"] Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.827679 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg2w7\" (UniqueName: \"kubernetes.io/projected/3637b659-50e8-4bd7-87a5-918a50b0e73e-kube-api-access-hg2w7\") pod \"route-controller-manager-84d5f88f56-kmn2q\" (UID: \"3637b659-50e8-4bd7-87a5-918a50b0e73e\") " pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.930510 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7caf2887-b32b-4d3d-9c93-a116076dde2c" path="/var/lib/kubelet/pods/7caf2887-b32b-4d3d-9c93-a116076dde2c/volumes" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.931330 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9821b1bc-27fb-4720-afce-16072b779105" path="/var/lib/kubelet/pods/9821b1bc-27fb-4720-afce-16072b779105/volumes" Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.951564 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-f9rdk"] Dec 02 10:24:52 crc kubenswrapper[4679]: I1202 10:24:52.958530 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.232407 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q"] Dec 02 10:24:53 crc kubenswrapper[4679]: W1202 10:24:53.242653 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3637b659_50e8_4bd7_87a5_918a50b0e73e.slice/crio-88b0858d906e141897b48dc3264116535de8d814340f0ae40cfb6fbb508ce16a WatchSource:0}: Error finding container 88b0858d906e141897b48dc3264116535de8d814340f0ae40cfb6fbb508ce16a: Status 404 returned error can't find the container with id 88b0858d906e141897b48dc3264116535de8d814340f0ae40cfb6fbb508ce16a Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.760775 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" event={"ID":"3637b659-50e8-4bd7-87a5-918a50b0e73e","Type":"ContainerStarted","Data":"5fab5ba38b1ad90bd05591b51a4b0ca471fea5e20a77445faf55269d83a22c80"} Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.761114 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" event={"ID":"3637b659-50e8-4bd7-87a5-918a50b0e73e","Type":"ContainerStarted","Data":"88b0858d906e141897b48dc3264116535de8d814340f0ae40cfb6fbb508ce16a"} Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.763044 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.764552 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" event={"ID":"5061ae44-8130-4aaf-8258-1bed986489af","Type":"ContainerStarted","Data":"56b379df39c03b097080e6955031331c5e5981686c4e3ffe30c05b0ad5f81785"} Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.764615 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" event={"ID":"5061ae44-8130-4aaf-8258-1bed986489af","Type":"ContainerStarted","Data":"48e88ed9a8b19ea4a2037d224120223ddad2aacfd80db51ba39d4f75670a0599"} Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.764822 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.770482 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.810775 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" podStartSLOduration=3.810758369 podStartE2EDuration="3.810758369s" podCreationTimestamp="2025-12-02 10:24:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:24:53.785834371 +0000 UTC m=+347.115973251" watchObservedRunningTime="2025-12-02 10:24:53.810758369 +0000 UTC m=+347.140897229" Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.811655 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" podStartSLOduration=1.811648865 podStartE2EDuration="1.811648865s" podCreationTimestamp="2025-12-02 10:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:24:53.80838855 +0000 UTC m=+347.138527400" watchObservedRunningTime="2025-12-02 10:24:53.811648865 +0000 UTC m=+347.141787725" Dec 02 10:24:53 crc kubenswrapper[4679]: I1202 10:24:53.950021 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84d5f88f56-kmn2q" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.700738 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f82lc"] Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.702180 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.710077 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f82lc"] Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842407 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f8c1d416-4446-4252-b525-31866ce579b4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842453 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842478 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-bound-sa-token\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842497 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8c1d416-4446-4252-b525-31866ce579b4-trusted-ca\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842656 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-registry-tls\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842704 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbxp6\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-kube-api-access-sbxp6\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842878 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f8c1d416-4446-4252-b525-31866ce579b4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.842953 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f8c1d416-4446-4252-b525-31866ce579b4-registry-certificates\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.862622 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944355 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f8c1d416-4446-4252-b525-31866ce579b4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944413 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f8c1d416-4446-4252-b525-31866ce579b4-registry-certificates\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944444 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f8c1d416-4446-4252-b525-31866ce579b4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944471 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-bound-sa-token\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944491 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8c1d416-4446-4252-b525-31866ce579b4-trusted-ca\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944538 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-registry-tls\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.944569 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbxp6\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-kube-api-access-sbxp6\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.945086 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f8c1d416-4446-4252-b525-31866ce579b4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.945716 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f8c1d416-4446-4252-b525-31866ce579b4-registry-certificates\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.946342 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8c1d416-4446-4252-b525-31866ce579b4-trusted-ca\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.951381 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f8c1d416-4446-4252-b525-31866ce579b4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.951799 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-registry-tls\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.967869 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbxp6\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-kube-api-access-sbxp6\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:10 crc kubenswrapper[4679]: I1202 10:25:10.974854 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8c1d416-4446-4252-b525-31866ce579b4-bound-sa-token\") pod \"image-registry-66df7c8f76-f82lc\" (UID: \"f8c1d416-4446-4252-b525-31866ce579b4\") " pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:11 crc kubenswrapper[4679]: I1202 10:25:11.021752 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:11 crc kubenswrapper[4679]: I1202 10:25:11.434149 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-f82lc"] Dec 02 10:25:11 crc kubenswrapper[4679]: W1202 10:25:11.445441 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8c1d416_4446_4252_b525_31866ce579b4.slice/crio-1a104b5cd9b9ab832bc3af032cece7caf18da153d1177e16f44444ec95c1efea WatchSource:0}: Error finding container 1a104b5cd9b9ab832bc3af032cece7caf18da153d1177e16f44444ec95c1efea: Status 404 returned error can't find the container with id 1a104b5cd9b9ab832bc3af032cece7caf18da153d1177e16f44444ec95c1efea Dec 02 10:25:11 crc kubenswrapper[4679]: I1202 10:25:11.875124 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" event={"ID":"f8c1d416-4446-4252-b525-31866ce579b4","Type":"ContainerStarted","Data":"f839f054082a17d29b4ed0a86505479e876e4492e5e3489813f19649b592b455"} Dec 02 10:25:11 crc kubenswrapper[4679]: I1202 10:25:11.875486 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" event={"ID":"f8c1d416-4446-4252-b525-31866ce579b4","Type":"ContainerStarted","Data":"1a104b5cd9b9ab832bc3af032cece7caf18da153d1177e16f44444ec95c1efea"} Dec 02 10:25:11 crc kubenswrapper[4679]: I1202 10:25:11.875781 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:11 crc kubenswrapper[4679]: I1202 10:25:11.893625 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" podStartSLOduration=1.893609031 podStartE2EDuration="1.893609031s" podCreationTimestamp="2025-12-02 10:25:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:25:11.892565271 +0000 UTC m=+365.222704141" watchObservedRunningTime="2025-12-02 10:25:11.893609031 +0000 UTC m=+365.223747891" Dec 02 10:25:16 crc kubenswrapper[4679]: I1202 10:25:16.932162 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:25:16 crc kubenswrapper[4679]: I1202 10:25:16.932228 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.619346 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hjp6m"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.620203 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hjp6m" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="registry-server" containerID="cri-o://7fb5a702b248b59c0339c642ffd4fa888329f5b181d4402c2e1f041cb7c47ad8" gracePeriod=30 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.623548 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vr2r9"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.623859 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vr2r9" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="registry-server" containerID="cri-o://89171d85207d29fb63658d2acd7ae4d962339fb3b9484fc692fc136200ad18f9" gracePeriod=30 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.636737 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hmbk"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.636989 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" containerID="cri-o://5fab1dc017b001583cc28843c17075f09fb8940ef13a6c8743f503d8b05274ba" gracePeriod=30 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.640496 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmqnd"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.640761 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vmqnd" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="registry-server" containerID="cri-o://4b65718c7eb0d676bdea3bacafa0759ee0a10fc60ef7f082a16d2ec2b85a0f9b" gracePeriod=30 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.649939 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-72flf"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.650831 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.657813 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-76kbb"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.658016 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-76kbb" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="registry-server" containerID="cri-o://3eb5d76f591f2e22cd217e976963c9c4bf13cb4dd04f3d14b87cdf2441e3b5b0" gracePeriod=30 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.658949 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-72flf"] Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.714929 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/de76d395-f3a6-474d-8ab7-f31c46710a0d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.715057 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de76d395-f3a6-474d-8ab7-f31c46710a0d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.715113 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kglfd\" (UniqueName: \"kubernetes.io/projected/de76d395-f3a6-474d-8ab7-f31c46710a0d-kube-api-access-kglfd\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.817129 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de76d395-f3a6-474d-8ab7-f31c46710a0d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.817401 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kglfd\" (UniqueName: \"kubernetes.io/projected/de76d395-f3a6-474d-8ab7-f31c46710a0d-kube-api-access-kglfd\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.817762 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/de76d395-f3a6-474d-8ab7-f31c46710a0d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.818807 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de76d395-f3a6-474d-8ab7-f31c46710a0d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.825896 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/de76d395-f3a6-474d-8ab7-f31c46710a0d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.840246 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kglfd\" (UniqueName: \"kubernetes.io/projected/de76d395-f3a6-474d-8ab7-f31c46710a0d-kube-api-access-kglfd\") pod \"marketplace-operator-79b997595-72flf\" (UID: \"de76d395-f3a6-474d-8ab7-f31c46710a0d\") " pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.961563 4679 generic.go:334] "Generic (PLEG): container finished" podID="dae67faf-d495-4e19-9c86-863362c49423" containerID="3eb5d76f591f2e22cd217e976963c9c4bf13cb4dd04f3d14b87cdf2441e3b5b0" exitCode=0 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.961646 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerDied","Data":"3eb5d76f591f2e22cd217e976963c9c4bf13cb4dd04f3d14b87cdf2441e3b5b0"} Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.964126 4679 generic.go:334] "Generic (PLEG): container finished" podID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerID="5fab1dc017b001583cc28843c17075f09fb8940ef13a6c8743f503d8b05274ba" exitCode=0 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.964204 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" event={"ID":"a180dc8a-e1e6-4da9-bf62-e973ef9d4611","Type":"ContainerDied","Data":"5fab1dc017b001583cc28843c17075f09fb8940ef13a6c8743f503d8b05274ba"} Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.964245 4679 scope.go:117] "RemoveContainer" containerID="f09cea1eb410b7707a4fe11c8b8df1052a5fd0bfabcc95d71f61365de9beae3a" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.970762 4679 generic.go:334] "Generic (PLEG): container finished" podID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerID="4b65718c7eb0d676bdea3bacafa0759ee0a10fc60ef7f082a16d2ec2b85a0f9b" exitCode=0 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.970930 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmqnd" event={"ID":"3dfcd027-05cf-48bc-8f61-f14a8d2a611f","Type":"ContainerDied","Data":"4b65718c7eb0d676bdea3bacafa0759ee0a10fc60ef7f082a16d2ec2b85a0f9b"} Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.977671 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.987757 4679 generic.go:334] "Generic (PLEG): container finished" podID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerID="7fb5a702b248b59c0339c642ffd4fa888329f5b181d4402c2e1f041cb7c47ad8" exitCode=0 Dec 02 10:25:27 crc kubenswrapper[4679]: I1202 10:25:27.987864 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hjp6m" event={"ID":"1a86b0c7-517b-465f-bbe5-d957b687a38d","Type":"ContainerDied","Data":"7fb5a702b248b59c0339c642ffd4fa888329f5b181d4402c2e1f041cb7c47ad8"} Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:27.998053 4679 generic.go:334] "Generic (PLEG): container finished" podID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerID="89171d85207d29fb63658d2acd7ae4d962339fb3b9484fc692fc136200ad18f9" exitCode=0 Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:27.998094 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2r9" event={"ID":"91618844-4970-41c8-ba39-ed76f10a7ff5","Type":"ContainerDied","Data":"89171d85207d29fb63658d2acd7ae4d962339fb3b9484fc692fc136200ad18f9"} Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.102213 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.123430 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-utilities\") pod \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.123528 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-catalog-content\") pod \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.123711 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmpmx\" (UniqueName: \"kubernetes.io/projected/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-kube-api-access-mmpmx\") pod \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\" (UID: \"3dfcd027-05cf-48bc-8f61-f14a8d2a611f\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.125009 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-utilities" (OuterVolumeSpecName: "utilities") pod "3dfcd027-05cf-48bc-8f61-f14a8d2a611f" (UID: "3dfcd027-05cf-48bc-8f61-f14a8d2a611f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.135501 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-kube-api-access-mmpmx" (OuterVolumeSpecName: "kube-api-access-mmpmx") pod "3dfcd027-05cf-48bc-8f61-f14a8d2a611f" (UID: "3dfcd027-05cf-48bc-8f61-f14a8d2a611f"). InnerVolumeSpecName "kube-api-access-mmpmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.169405 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dfcd027-05cf-48bc-8f61-f14a8d2a611f" (UID: "3dfcd027-05cf-48bc-8f61-f14a8d2a611f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.226197 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.226230 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.226270 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmpmx\" (UniqueName: \"kubernetes.io/projected/3dfcd027-05cf-48bc-8f61-f14a8d2a611f-kube-api-access-mmpmx\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.257495 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.279642 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.283404 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.295116 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327090 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-trusted-ca\") pod \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327139 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqgzw\" (UniqueName: \"kubernetes.io/projected/1a86b0c7-517b-465f-bbe5-d957b687a38d-kube-api-access-jqgzw\") pod \"1a86b0c7-517b-465f-bbe5-d957b687a38d\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327173 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-utilities\") pod \"91618844-4970-41c8-ba39-ed76f10a7ff5\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327200 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67pln\" (UniqueName: \"kubernetes.io/projected/dae67faf-d495-4e19-9c86-863362c49423-kube-api-access-67pln\") pod \"dae67faf-d495-4e19-9c86-863362c49423\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327241 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-catalog-content\") pod \"dae67faf-d495-4e19-9c86-863362c49423\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327291 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-catalog-content\") pod \"1a86b0c7-517b-465f-bbe5-d957b687a38d\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327336 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhb88\" (UniqueName: \"kubernetes.io/projected/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-kube-api-access-nhb88\") pod \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327368 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-operator-metrics\") pod \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\" (UID: \"a180dc8a-e1e6-4da9-bf62-e973ef9d4611\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327391 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-utilities\") pod \"1a86b0c7-517b-465f-bbe5-d957b687a38d\" (UID: \"1a86b0c7-517b-465f-bbe5-d957b687a38d\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327423 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99btx\" (UniqueName: \"kubernetes.io/projected/91618844-4970-41c8-ba39-ed76f10a7ff5-kube-api-access-99btx\") pod \"91618844-4970-41c8-ba39-ed76f10a7ff5\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327447 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-utilities\") pod \"dae67faf-d495-4e19-9c86-863362c49423\" (UID: \"dae67faf-d495-4e19-9c86-863362c49423\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.327485 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-catalog-content\") pod \"91618844-4970-41c8-ba39-ed76f10a7ff5\" (UID: \"91618844-4970-41c8-ba39-ed76f10a7ff5\") " Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.330942 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a180dc8a-e1e6-4da9-bf62-e973ef9d4611" (UID: "a180dc8a-e1e6-4da9-bf62-e973ef9d4611"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.335112 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-utilities" (OuterVolumeSpecName: "utilities") pod "1a86b0c7-517b-465f-bbe5-d957b687a38d" (UID: "1a86b0c7-517b-465f-bbe5-d957b687a38d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.335768 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-utilities" (OuterVolumeSpecName: "utilities") pod "dae67faf-d495-4e19-9c86-863362c49423" (UID: "dae67faf-d495-4e19-9c86-863362c49423"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.336904 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-utilities" (OuterVolumeSpecName: "utilities") pod "91618844-4970-41c8-ba39-ed76f10a7ff5" (UID: "91618844-4970-41c8-ba39-ed76f10a7ff5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.338967 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dae67faf-d495-4e19-9c86-863362c49423-kube-api-access-67pln" (OuterVolumeSpecName: "kube-api-access-67pln") pod "dae67faf-d495-4e19-9c86-863362c49423" (UID: "dae67faf-d495-4e19-9c86-863362c49423"). InnerVolumeSpecName "kube-api-access-67pln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.339265 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a86b0c7-517b-465f-bbe5-d957b687a38d-kube-api-access-jqgzw" (OuterVolumeSpecName: "kube-api-access-jqgzw") pod "1a86b0c7-517b-465f-bbe5-d957b687a38d" (UID: "1a86b0c7-517b-465f-bbe5-d957b687a38d"). InnerVolumeSpecName "kube-api-access-jqgzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.340227 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a180dc8a-e1e6-4da9-bf62-e973ef9d4611" (UID: "a180dc8a-e1e6-4da9-bf62-e973ef9d4611"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.341675 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91618844-4970-41c8-ba39-ed76f10a7ff5-kube-api-access-99btx" (OuterVolumeSpecName: "kube-api-access-99btx") pod "91618844-4970-41c8-ba39-ed76f10a7ff5" (UID: "91618844-4970-41c8-ba39-ed76f10a7ff5"). InnerVolumeSpecName "kube-api-access-99btx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.342465 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-kube-api-access-nhb88" (OuterVolumeSpecName: "kube-api-access-nhb88") pod "a180dc8a-e1e6-4da9-bf62-e973ef9d4611" (UID: "a180dc8a-e1e6-4da9-bf62-e973ef9d4611"). InnerVolumeSpecName "kube-api-access-nhb88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.384038 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91618844-4970-41c8-ba39-ed76f10a7ff5" (UID: "91618844-4970-41c8-ba39-ed76f10a7ff5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.396982 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a86b0c7-517b-465f-bbe5-d957b687a38d" (UID: "1a86b0c7-517b-465f-bbe5-d957b687a38d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431031 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99btx\" (UniqueName: \"kubernetes.io/projected/91618844-4970-41c8-ba39-ed76f10a7ff5-kube-api-access-99btx\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431072 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431087 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431099 4679 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431110 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqgzw\" (UniqueName: \"kubernetes.io/projected/1a86b0c7-517b-465f-bbe5-d957b687a38d-kube-api-access-jqgzw\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431121 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91618844-4970-41c8-ba39-ed76f10a7ff5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431132 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67pln\" (UniqueName: \"kubernetes.io/projected/dae67faf-d495-4e19-9c86-863362c49423-kube-api-access-67pln\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431143 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431154 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhb88\" (UniqueName: \"kubernetes.io/projected/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-kube-api-access-nhb88\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431166 4679 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a180dc8a-e1e6-4da9-bf62-e973ef9d4611-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.431178 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a86b0c7-517b-465f-bbe5-d957b687a38d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.471878 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dae67faf-d495-4e19-9c86-863362c49423" (UID: "dae67faf-d495-4e19-9c86-863362c49423"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.476384 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-72flf"] Dec 02 10:25:28 crc kubenswrapper[4679]: I1202 10:25:28.532093 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae67faf-d495-4e19-9c86-863362c49423-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.005859 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hjp6m" event={"ID":"1a86b0c7-517b-465f-bbe5-d957b687a38d","Type":"ContainerDied","Data":"386ee80c7a329decc9b01889192ccda60288b164bfb2cd52f52cd45ccdaa8340"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.005934 4679 scope.go:117] "RemoveContainer" containerID="7fb5a702b248b59c0339c642ffd4fa888329f5b181d4402c2e1f041cb7c47ad8" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.005880 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hjp6m" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.008611 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vr2r9" event={"ID":"91618844-4970-41c8-ba39-ed76f10a7ff5","Type":"ContainerDied","Data":"cd237ec8fc6e9d73195aed40baea086f5f60ca6145a329a2aed999ae9d124c6d"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.008846 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vr2r9" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.013008 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-76kbb" event={"ID":"dae67faf-d495-4e19-9c86-863362c49423","Type":"ContainerDied","Data":"5afe77130a795005eb328cd71825c9571d8420a7f908ed6834785f104c15d398"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.013036 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-76kbb" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.014844 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" event={"ID":"a180dc8a-e1e6-4da9-bf62-e973ef9d4611","Type":"ContainerDied","Data":"8442f2ad5edda06d137c820f12bbea689c4fd8bfddb44ff703b570b12fcfa808"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.014885 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4hmbk" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.018058 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmqnd" event={"ID":"3dfcd027-05cf-48bc-8f61-f14a8d2a611f","Type":"ContainerDied","Data":"5158e51d1e26c148ddcd37fb389a43d8ae12a3ab9db27bba0544f42cfcb1ef45"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.018157 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmqnd" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.020840 4679 scope.go:117] "RemoveContainer" containerID="25a3bc0a7188d284f8476625c3ea947b515951f5e89d61ba63097b753b98fa95" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.021660 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" event={"ID":"de76d395-f3a6-474d-8ab7-f31c46710a0d","Type":"ContainerStarted","Data":"f586e223b1dfd6d5e14469b044870d112190fdc87df68ded833e5cbc4715a492"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.021716 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" event={"ID":"de76d395-f3a6-474d-8ab7-f31c46710a0d","Type":"ContainerStarted","Data":"9487b3101477330fc1705992f71d48679f5d214b42e10470e7c91659cb2a91c0"} Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.022132 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.028502 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.045403 4679 scope.go:117] "RemoveContainer" containerID="0fc18ae0706582676d8211e94c8563222ed6528ee37e97a3f50267f62276b5fd" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.045548 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vr2r9"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.048883 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vr2r9"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.053231 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hjp6m"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.056909 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hjp6m"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.060071 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-76kbb"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.066078 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-76kbb"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.066372 4679 scope.go:117] "RemoveContainer" containerID="89171d85207d29fb63658d2acd7ae4d962339fb3b9484fc692fc136200ad18f9" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.073115 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-72flf" podStartSLOduration=2.073095087 podStartE2EDuration="2.073095087s" podCreationTimestamp="2025-12-02 10:25:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:25:29.064874977 +0000 UTC m=+382.395013837" watchObservedRunningTime="2025-12-02 10:25:29.073095087 +0000 UTC m=+382.403233947" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.078846 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmqnd"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.084291 4679 scope.go:117] "RemoveContainer" containerID="d804594ed1b5f3aba71e2532c9a79b1fb58a80df90f86d872d08313ac87c2d93" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.086889 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmqnd"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.091426 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hmbk"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.094642 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4hmbk"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.104719 4679 scope.go:117] "RemoveContainer" containerID="329ba386a82361670bbabdbce0032f3226d52ffedddb89c4272179cf56aac47c" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.117984 4679 scope.go:117] "RemoveContainer" containerID="3eb5d76f591f2e22cd217e976963c9c4bf13cb4dd04f3d14b87cdf2441e3b5b0" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.133993 4679 scope.go:117] "RemoveContainer" containerID="7c9695d9473b5f50bbec8b1ae164aca801823e2dca043be1ca2c064d3d41aef8" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.147298 4679 scope.go:117] "RemoveContainer" containerID="c9e1947fb9faf037921636b1447729913f8722ca7e188581921020c1f2e750d6" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.162511 4679 scope.go:117] "RemoveContainer" containerID="5fab1dc017b001583cc28843c17075f09fb8940ef13a6c8743f503d8b05274ba" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.176571 4679 scope.go:117] "RemoveContainer" containerID="4b65718c7eb0d676bdea3bacafa0759ee0a10fc60ef7f082a16d2ec2b85a0f9b" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.189351 4679 scope.go:117] "RemoveContainer" containerID="9552f093fa92daf277a9923c688c4b9f472b8ec7876e72dbf12e314c40965f03" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.205215 4679 scope.go:117] "RemoveContainer" containerID="fd878ff823515af7ce151a0631d27abdb64dc7abe8e5368211735726fce21233" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826206 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dg7kk"] Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826423 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826433 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826444 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826450 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826458 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826464 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826472 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826478 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826486 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826491 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826499 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826504 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826512 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826518 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826523 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826529 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="extract-content" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826538 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826543 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826551 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826556 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826563 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826569 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826577 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826582 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="extract-utilities" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826591 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826597 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826671 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826680 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826687 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826698 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826705 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826712 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="dae67faf-d495-4e19-9c86-863362c49423" containerName="registry-server" Dec 02 10:25:29 crc kubenswrapper[4679]: E1202 10:25:29.826794 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.826801 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" containerName="marketplace-operator" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.827712 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.830845 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.835097 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg7kk"] Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.952456 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-catalog-content\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.952843 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggx5d\" (UniqueName: \"kubernetes.io/projected/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-kube-api-access-ggx5d\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:29 crc kubenswrapper[4679]: I1202 10:25:29.953082 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-utilities\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.029875 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dk6hk"] Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.032655 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.036493 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.041572 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk6hk"] Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.055175 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-catalog-content\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.055216 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggx5d\" (UniqueName: \"kubernetes.io/projected/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-kube-api-access-ggx5d\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.055289 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-utilities\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.055719 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-catalog-content\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.055911 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-utilities\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.073467 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggx5d\" (UniqueName: \"kubernetes.io/projected/1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba-kube-api-access-ggx5d\") pod \"redhat-marketplace-dg7kk\" (UID: \"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba\") " pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.144902 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.156114 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33efe285-5788-489a-98b5-c6cce7f2738c-catalog-content\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.156209 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgp4w\" (UniqueName: \"kubernetes.io/projected/33efe285-5788-489a-98b5-c6cce7f2738c-kube-api-access-xgp4w\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.156258 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33efe285-5788-489a-98b5-c6cce7f2738c-utilities\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.257240 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgp4w\" (UniqueName: \"kubernetes.io/projected/33efe285-5788-489a-98b5-c6cce7f2738c-kube-api-access-xgp4w\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.257623 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33efe285-5788-489a-98b5-c6cce7f2738c-utilities\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.257671 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33efe285-5788-489a-98b5-c6cce7f2738c-catalog-content\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.258608 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33efe285-5788-489a-98b5-c6cce7f2738c-catalog-content\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.260085 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33efe285-5788-489a-98b5-c6cce7f2738c-utilities\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.280911 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgp4w\" (UniqueName: \"kubernetes.io/projected/33efe285-5788-489a-98b5-c6cce7f2738c-kube-api-access-xgp4w\") pod \"certified-operators-dk6hk\" (UID: \"33efe285-5788-489a-98b5-c6cce7f2738c\") " pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.353973 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.544678 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg7kk"] Dec 02 10:25:30 crc kubenswrapper[4679]: W1202 10:25:30.546617 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f4b93fd_d2b3_4f28_aff5_1a4dd8d79cba.slice/crio-a8e3170a8a49efb7ab0893f7aa246eabc33a2bddbcb1c6f5a210ff53fcc041ba WatchSource:0}: Error finding container a8e3170a8a49efb7ab0893f7aa246eabc33a2bddbcb1c6f5a210ff53fcc041ba: Status 404 returned error can't find the container with id a8e3170a8a49efb7ab0893f7aa246eabc33a2bddbcb1c6f5a210ff53fcc041ba Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.744780 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk6hk"] Dec 02 10:25:30 crc kubenswrapper[4679]: E1202 10:25:30.765791 4679 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f4b93fd_d2b3_4f28_aff5_1a4dd8d79cba.slice/crio-00372002e003e33ff55756c83ad534d62d2be82a734261ee05fbd0a1ed2561a1.scope\": RecentStats: unable to find data in memory cache]" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.807100 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-f9rdk"] Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.807288 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" podUID="5061ae44-8130-4aaf-8258-1bed986489af" containerName="controller-manager" containerID="cri-o://56b379df39c03b097080e6955031331c5e5981686c4e3ffe30c05b0ad5f81785" gracePeriod=30 Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.914200 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a86b0c7-517b-465f-bbe5-d957b687a38d" path="/var/lib/kubelet/pods/1a86b0c7-517b-465f-bbe5-d957b687a38d/volumes" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.914985 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dfcd027-05cf-48bc-8f61-f14a8d2a611f" path="/var/lib/kubelet/pods/3dfcd027-05cf-48bc-8f61-f14a8d2a611f/volumes" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.915531 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91618844-4970-41c8-ba39-ed76f10a7ff5" path="/var/lib/kubelet/pods/91618844-4970-41c8-ba39-ed76f10a7ff5/volumes" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.916555 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a180dc8a-e1e6-4da9-bf62-e973ef9d4611" path="/var/lib/kubelet/pods/a180dc8a-e1e6-4da9-bf62-e973ef9d4611/volumes" Dec 02 10:25:30 crc kubenswrapper[4679]: I1202 10:25:30.916980 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dae67faf-d495-4e19-9c86-863362c49423" path="/var/lib/kubelet/pods/dae67faf-d495-4e19-9c86-863362c49423/volumes" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.027501 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-f82lc" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.057874 4679 generic.go:334] "Generic (PLEG): container finished" podID="1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba" containerID="00372002e003e33ff55756c83ad534d62d2be82a734261ee05fbd0a1ed2561a1" exitCode=0 Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.057949 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg7kk" event={"ID":"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba","Type":"ContainerDied","Data":"00372002e003e33ff55756c83ad534d62d2be82a734261ee05fbd0a1ed2561a1"} Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.058018 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg7kk" event={"ID":"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba","Type":"ContainerStarted","Data":"a8e3170a8a49efb7ab0893f7aa246eabc33a2bddbcb1c6f5a210ff53fcc041ba"} Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.062033 4679 generic.go:334] "Generic (PLEG): container finished" podID="33efe285-5788-489a-98b5-c6cce7f2738c" containerID="5027f82c61111a5f6851aaf334e20e5d7933e967a4812c77c7f710b730cf2696" exitCode=0 Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.062096 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk6hk" event={"ID":"33efe285-5788-489a-98b5-c6cce7f2738c","Type":"ContainerDied","Data":"5027f82c61111a5f6851aaf334e20e5d7933e967a4812c77c7f710b730cf2696"} Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.062120 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk6hk" event={"ID":"33efe285-5788-489a-98b5-c6cce7f2738c","Type":"ContainerStarted","Data":"84fc2d313e23ec9a38a98019ddc4caff3a954b80a619e9536c804a05630f5a5e"} Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.072282 4679 generic.go:334] "Generic (PLEG): container finished" podID="5061ae44-8130-4aaf-8258-1bed986489af" containerID="56b379df39c03b097080e6955031331c5e5981686c4e3ffe30c05b0ad5f81785" exitCode=0 Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.072970 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" event={"ID":"5061ae44-8130-4aaf-8258-1bed986489af","Type":"ContainerDied","Data":"56b379df39c03b097080e6955031331c5e5981686c4e3ffe30c05b0ad5f81785"} Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.084740 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h2ljp"] Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.239222 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.291840 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-proxy-ca-bundles\") pod \"5061ae44-8130-4aaf-8258-1bed986489af\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.291889 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-client-ca\") pod \"5061ae44-8130-4aaf-8258-1bed986489af\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.291971 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-config\") pod \"5061ae44-8130-4aaf-8258-1bed986489af\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.292004 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwxw7\" (UniqueName: \"kubernetes.io/projected/5061ae44-8130-4aaf-8258-1bed986489af-kube-api-access-wwxw7\") pod \"5061ae44-8130-4aaf-8258-1bed986489af\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.292027 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5061ae44-8130-4aaf-8258-1bed986489af-serving-cert\") pod \"5061ae44-8130-4aaf-8258-1bed986489af\" (UID: \"5061ae44-8130-4aaf-8258-1bed986489af\") " Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.293728 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-client-ca" (OuterVolumeSpecName: "client-ca") pod "5061ae44-8130-4aaf-8258-1bed986489af" (UID: "5061ae44-8130-4aaf-8258-1bed986489af"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.294409 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-config" (OuterVolumeSpecName: "config") pod "5061ae44-8130-4aaf-8258-1bed986489af" (UID: "5061ae44-8130-4aaf-8258-1bed986489af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.295124 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5061ae44-8130-4aaf-8258-1bed986489af" (UID: "5061ae44-8130-4aaf-8258-1bed986489af"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.297028 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5061ae44-8130-4aaf-8258-1bed986489af-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5061ae44-8130-4aaf-8258-1bed986489af" (UID: "5061ae44-8130-4aaf-8258-1bed986489af"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.297030 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5061ae44-8130-4aaf-8258-1bed986489af-kube-api-access-wwxw7" (OuterVolumeSpecName: "kube-api-access-wwxw7") pod "5061ae44-8130-4aaf-8258-1bed986489af" (UID: "5061ae44-8130-4aaf-8258-1bed986489af"). InnerVolumeSpecName "kube-api-access-wwxw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.393671 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.393720 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwxw7\" (UniqueName: \"kubernetes.io/projected/5061ae44-8130-4aaf-8258-1bed986489af-kube-api-access-wwxw7\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.393737 4679 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5061ae44-8130-4aaf-8258-1bed986489af-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.393759 4679 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:31 crc kubenswrapper[4679]: I1202 10:25:31.393771 4679 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5061ae44-8130-4aaf-8258-1bed986489af-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.081327 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" event={"ID":"5061ae44-8130-4aaf-8258-1bed986489af","Type":"ContainerDied","Data":"48e88ed9a8b19ea4a2037d224120223ddad2aacfd80db51ba39d4f75670a0599"} Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.081723 4679 scope.go:117] "RemoveContainer" containerID="56b379df39c03b097080e6955031331c5e5981686c4e3ffe30c05b0ad5f81785" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.081363 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-f9rdk" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.084667 4679 generic.go:334] "Generic (PLEG): container finished" podID="1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba" containerID="c47a55749d1b89e4dc69230220aa362c8d75c2fcaa511bf74443fa28766ce8ed" exitCode=0 Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.084771 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg7kk" event={"ID":"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba","Type":"ContainerDied","Data":"c47a55749d1b89e4dc69230220aa362c8d75c2fcaa511bf74443fa28766ce8ed"} Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.129022 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-f9rdk"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.133218 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-f9rdk"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.221847 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z6xb5"] Dec 02 10:25:32 crc kubenswrapper[4679]: E1202 10:25:32.222105 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5061ae44-8130-4aaf-8258-1bed986489af" containerName="controller-manager" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.222121 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5061ae44-8130-4aaf-8258-1bed986489af" containerName="controller-manager" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.222225 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5061ae44-8130-4aaf-8258-1bed986489af" containerName="controller-manager" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.223072 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.227230 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.238374 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6xb5"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.305292 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e534b0e4-8964-48d5-8bad-734ec5d93a8b-catalog-content\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.305552 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq829\" (UniqueName: \"kubernetes.io/projected/e534b0e4-8964-48d5-8bad-734ec5d93a8b-kube-api-access-wq829\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.305774 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e534b0e4-8964-48d5-8bad-734ec5d93a8b-utilities\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.406741 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e534b0e4-8964-48d5-8bad-734ec5d93a8b-catalog-content\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.406822 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq829\" (UniqueName: \"kubernetes.io/projected/e534b0e4-8964-48d5-8bad-734ec5d93a8b-kube-api-access-wq829\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.406874 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e534b0e4-8964-48d5-8bad-734ec5d93a8b-utilities\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.407661 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e534b0e4-8964-48d5-8bad-734ec5d93a8b-utilities\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.408564 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e534b0e4-8964-48d5-8bad-734ec5d93a8b-catalog-content\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.427513 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq829\" (UniqueName: \"kubernetes.io/projected/e534b0e4-8964-48d5-8bad-734ec5d93a8b-kube-api-access-wq829\") pod \"redhat-operators-z6xb5\" (UID: \"e534b0e4-8964-48d5-8bad-734ec5d93a8b\") " pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.429242 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-76xkq"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.430913 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.432973 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.440409 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76xkq"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.508797 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-catalog-content\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.508969 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcb82\" (UniqueName: \"kubernetes.io/projected/5610cf13-1e22-4bb5-a5da-3b67207aef40-kube-api-access-xcb82\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.509015 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-utilities\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.578039 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.610840 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcb82\" (UniqueName: \"kubernetes.io/projected/5610cf13-1e22-4bb5-a5da-3b67207aef40-kube-api-access-xcb82\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.610892 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-utilities\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.610940 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-catalog-content\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.611540 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-utilities\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.611681 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-catalog-content\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.638989 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcb82\" (UniqueName: \"kubernetes.io/projected/5610cf13-1e22-4bb5-a5da-3b67207aef40-kube-api-access-xcb82\") pod \"community-operators-76xkq\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.672987 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.674366 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.680899 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.680988 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.682165 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.682576 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.683883 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.688072 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.688207 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.691136 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn"] Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.780603 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.813998 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-config\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.814378 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-proxy-ca-bundles\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.814559 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5swg\" (UniqueName: \"kubernetes.io/projected/5544a554-b2c2-41ff-a36a-5279ee0d4546-kube-api-access-f5swg\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.814634 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5544a554-b2c2-41ff-a36a-5279ee0d4546-serving-cert\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.814697 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-client-ca\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.916084 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5swg\" (UniqueName: \"kubernetes.io/projected/5544a554-b2c2-41ff-a36a-5279ee0d4546-kube-api-access-f5swg\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.916141 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5544a554-b2c2-41ff-a36a-5279ee0d4546-serving-cert\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.916173 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-client-ca\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.916192 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-config\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.916251 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-proxy-ca-bundles\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.917008 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5061ae44-8130-4aaf-8258-1bed986489af" path="/var/lib/kubelet/pods/5061ae44-8130-4aaf-8258-1bed986489af/volumes" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.917610 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-client-ca\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.918069 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-proxy-ca-bundles\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.918424 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5544a554-b2c2-41ff-a36a-5279ee0d4546-config\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.924757 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5544a554-b2c2-41ff-a36a-5279ee0d4546-serving-cert\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.939059 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5swg\" (UniqueName: \"kubernetes.io/projected/5544a554-b2c2-41ff-a36a-5279ee0d4546-kube-api-access-f5swg\") pod \"controller-manager-5d7c6748bd-6nnqn\" (UID: \"5544a554-b2c2-41ff-a36a-5279ee0d4546\") " pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:32 crc kubenswrapper[4679]: I1202 10:25:32.979766 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76xkq"] Dec 02 10:25:32 crc kubenswrapper[4679]: W1202 10:25:32.985238 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5610cf13_1e22_4bb5_a5da_3b67207aef40.slice/crio-c7d8dc0c700c9d33ef1898f99133c27d41257118164af4b9b5aa22d6e1e9c2f2 WatchSource:0}: Error finding container c7d8dc0c700c9d33ef1898f99133c27d41257118164af4b9b5aa22d6e1e9c2f2: Status 404 returned error can't find the container with id c7d8dc0c700c9d33ef1898f99133c27d41257118164af4b9b5aa22d6e1e9c2f2 Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.040335 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.057298 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6xb5"] Dec 02 10:25:33 crc kubenswrapper[4679]: W1202 10:25:33.066076 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode534b0e4_8964_48d5_8bad_734ec5d93a8b.slice/crio-712c612bc50824d937a00f7fbc96e95829b236a70919462060d64428fbdd9af8 WatchSource:0}: Error finding container 712c612bc50824d937a00f7fbc96e95829b236a70919462060d64428fbdd9af8: Status 404 returned error can't find the container with id 712c612bc50824d937a00f7fbc96e95829b236a70919462060d64428fbdd9af8 Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.101161 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg7kk" event={"ID":"1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba","Type":"ContainerStarted","Data":"a77a6093e8ed8eaad5c15e29197deb7e008806f1781289c9a0422d07d5ec7f51"} Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.104752 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerStarted","Data":"f5f7c2604ac78f4621b37ef7fec64c1841a2d5aa617fd0042a959fddb22c6b2f"} Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.104778 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerStarted","Data":"c7d8dc0c700c9d33ef1898f99133c27d41257118164af4b9b5aa22d6e1e9c2f2"} Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.107677 4679 generic.go:334] "Generic (PLEG): container finished" podID="33efe285-5788-489a-98b5-c6cce7f2738c" containerID="1bdc194a61c563fad9dd18191ca9124eb8d7f19f4827a3e5d5d83af634bc1319" exitCode=0 Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.107782 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk6hk" event={"ID":"33efe285-5788-489a-98b5-c6cce7f2738c","Type":"ContainerDied","Data":"1bdc194a61c563fad9dd18191ca9124eb8d7f19f4827a3e5d5d83af634bc1319"} Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.110933 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6xb5" event={"ID":"e534b0e4-8964-48d5-8bad-734ec5d93a8b","Type":"ContainerStarted","Data":"712c612bc50824d937a00f7fbc96e95829b236a70919462060d64428fbdd9af8"} Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.123868 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dg7kk" podStartSLOduration=2.550293353 podStartE2EDuration="4.123854773s" podCreationTimestamp="2025-12-02 10:25:29 +0000 UTC" firstStartedPulling="2025-12-02 10:25:31.059222223 +0000 UTC m=+384.389361083" lastFinishedPulling="2025-12-02 10:25:32.632783633 +0000 UTC m=+385.962922503" observedRunningTime="2025-12-02 10:25:33.122579626 +0000 UTC m=+386.452718486" watchObservedRunningTime="2025-12-02 10:25:33.123854773 +0000 UTC m=+386.453993633" Dec 02 10:25:33 crc kubenswrapper[4679]: I1202 10:25:33.266961 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn"] Dec 02 10:25:33 crc kubenswrapper[4679]: W1202 10:25:33.318598 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5544a554_b2c2_41ff_a36a_5279ee0d4546.slice/crio-6480d82f63c2f07975c7ecec7ba20a9333b909a182769a7eca0f97605bb58338 WatchSource:0}: Error finding container 6480d82f63c2f07975c7ecec7ba20a9333b909a182769a7eca0f97605bb58338: Status 404 returned error can't find the container with id 6480d82f63c2f07975c7ecec7ba20a9333b909a182769a7eca0f97605bb58338 Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.117896 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" event={"ID":"5544a554-b2c2-41ff-a36a-5279ee0d4546","Type":"ContainerStarted","Data":"ccf7aceb30644a2a02bebfd741df5ff3159ed8039ad34ab44dbe38371bac5e3d"} Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.118402 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" event={"ID":"5544a554-b2c2-41ff-a36a-5279ee0d4546","Type":"ContainerStarted","Data":"6480d82f63c2f07975c7ecec7ba20a9333b909a182769a7eca0f97605bb58338"} Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.118424 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.119377 4679 generic.go:334] "Generic (PLEG): container finished" podID="e534b0e4-8964-48d5-8bad-734ec5d93a8b" containerID="7f3d3b739995fc52355d5baf182eeeacfb91ea15b11b4ed8d26a7423e2204d19" exitCode=0 Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.119459 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6xb5" event={"ID":"e534b0e4-8964-48d5-8bad-734ec5d93a8b","Type":"ContainerDied","Data":"7f3d3b739995fc52355d5baf182eeeacfb91ea15b11b4ed8d26a7423e2204d19"} Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.121918 4679 generic.go:334] "Generic (PLEG): container finished" podID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerID="f5f7c2604ac78f4621b37ef7fec64c1841a2d5aa617fd0042a959fddb22c6b2f" exitCode=0 Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.121986 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerDied","Data":"f5f7c2604ac78f4621b37ef7fec64c1841a2d5aa617fd0042a959fddb22c6b2f"} Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.122159 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.126882 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk6hk" event={"ID":"33efe285-5788-489a-98b5-c6cce7f2738c","Type":"ContainerStarted","Data":"de8696913d1afb2a58b516750eaedce6f02b22a9964cfd4a598f0604c0993de1"} Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.135868 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d7c6748bd-6nnqn" podStartSLOduration=4.135850814 podStartE2EDuration="4.135850814s" podCreationTimestamp="2025-12-02 10:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:25:34.131645701 +0000 UTC m=+387.461784571" watchObservedRunningTime="2025-12-02 10:25:34.135850814 +0000 UTC m=+387.465989674" Dec 02 10:25:34 crc kubenswrapper[4679]: I1202 10:25:34.209944 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dk6hk" podStartSLOduration=1.723212044 podStartE2EDuration="4.209926698s" podCreationTimestamp="2025-12-02 10:25:30 +0000 UTC" firstStartedPulling="2025-12-02 10:25:31.063676513 +0000 UTC m=+384.393815373" lastFinishedPulling="2025-12-02 10:25:33.550391167 +0000 UTC m=+386.880530027" observedRunningTime="2025-12-02 10:25:34.207992341 +0000 UTC m=+387.538131201" watchObservedRunningTime="2025-12-02 10:25:34.209926698 +0000 UTC m=+387.540065558" Dec 02 10:25:36 crc kubenswrapper[4679]: I1202 10:25:36.139031 4679 generic.go:334] "Generic (PLEG): container finished" podID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerID="46370d8c7e8e7dc32df22aed0891d3fc00c62705886a09f6f53b5c4583ebbcdf" exitCode=0 Dec 02 10:25:36 crc kubenswrapper[4679]: I1202 10:25:36.139228 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerDied","Data":"46370d8c7e8e7dc32df22aed0891d3fc00c62705886a09f6f53b5c4583ebbcdf"} Dec 02 10:25:36 crc kubenswrapper[4679]: I1202 10:25:36.141610 4679 generic.go:334] "Generic (PLEG): container finished" podID="e534b0e4-8964-48d5-8bad-734ec5d93a8b" containerID="cbcc4b723d21262874a894aa4e5583193e950ac7c37bd7a474adbb4d5beb37ae" exitCode=0 Dec 02 10:25:36 crc kubenswrapper[4679]: I1202 10:25:36.142050 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6xb5" event={"ID":"e534b0e4-8964-48d5-8bad-734ec5d93a8b","Type":"ContainerDied","Data":"cbcc4b723d21262874a894aa4e5583193e950ac7c37bd7a474adbb4d5beb37ae"} Dec 02 10:25:38 crc kubenswrapper[4679]: I1202 10:25:38.154164 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerStarted","Data":"2eece58ddecd8fdf638aa95f1cef9562ab4dc6d6253b12652d517cd54500cfb4"} Dec 02 10:25:38 crc kubenswrapper[4679]: I1202 10:25:38.157504 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6xb5" event={"ID":"e534b0e4-8964-48d5-8bad-734ec5d93a8b","Type":"ContainerStarted","Data":"86ab902eb94f5dcd0688d045d00457c1153b795de87fb465316521b125b291d6"} Dec 02 10:25:38 crc kubenswrapper[4679]: I1202 10:25:38.173334 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-76xkq" podStartSLOduration=3.493064712 podStartE2EDuration="6.173318991s" podCreationTimestamp="2025-12-02 10:25:32 +0000 UTC" firstStartedPulling="2025-12-02 10:25:34.123468122 +0000 UTC m=+387.453606982" lastFinishedPulling="2025-12-02 10:25:36.803722391 +0000 UTC m=+390.133861261" observedRunningTime="2025-12-02 10:25:38.170346995 +0000 UTC m=+391.500485855" watchObservedRunningTime="2025-12-02 10:25:38.173318991 +0000 UTC m=+391.503457851" Dec 02 10:25:38 crc kubenswrapper[4679]: I1202 10:25:38.194022 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z6xb5" podStartSLOduration=3.7025296770000002 podStartE2EDuration="6.194005639s" podCreationTimestamp="2025-12-02 10:25:32 +0000 UTC" firstStartedPulling="2025-12-02 10:25:34.121474694 +0000 UTC m=+387.451613554" lastFinishedPulling="2025-12-02 10:25:36.612950656 +0000 UTC m=+389.943089516" observedRunningTime="2025-12-02 10:25:38.191699543 +0000 UTC m=+391.521838443" watchObservedRunningTime="2025-12-02 10:25:38.194005639 +0000 UTC m=+391.524144499" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.145747 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.149114 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.192116 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.233541 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dg7kk" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.354987 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.355054 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:40 crc kubenswrapper[4679]: I1202 10:25:40.395179 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:41 crc kubenswrapper[4679]: I1202 10:25:41.236833 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dk6hk" Dec 02 10:25:42 crc kubenswrapper[4679]: I1202 10:25:42.578785 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:42 crc kubenswrapper[4679]: I1202 10:25:42.579127 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:42 crc kubenswrapper[4679]: I1202 10:25:42.623571 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:42 crc kubenswrapper[4679]: I1202 10:25:42.781605 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:42 crc kubenswrapper[4679]: I1202 10:25:42.781655 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:42 crc kubenswrapper[4679]: I1202 10:25:42.819551 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:43 crc kubenswrapper[4679]: I1202 10:25:43.235016 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:25:43 crc kubenswrapper[4679]: I1202 10:25:43.237457 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z6xb5" Dec 02 10:25:46 crc kubenswrapper[4679]: I1202 10:25:46.931646 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:25:46 crc kubenswrapper[4679]: I1202 10:25:46.931696 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:25:56 crc kubenswrapper[4679]: I1202 10:25:56.121965 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" podUID="cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" containerName="registry" containerID="cri-o://7bb13bdcf48b735fe12f95e6ba5a297afa05925396a7c2f6b96ddb17d137e050" gracePeriod=30 Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.277298 4679 generic.go:334] "Generic (PLEG): container finished" podID="cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" containerID="7bb13bdcf48b735fe12f95e6ba5a297afa05925396a7c2f6b96ddb17d137e050" exitCode=0 Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.277583 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" event={"ID":"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26","Type":"ContainerDied","Data":"7bb13bdcf48b735fe12f95e6ba5a297afa05925396a7c2f6b96ddb17d137e050"} Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.435807 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469432 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-bound-sa-token\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469650 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469674 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-trusted-ca\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469695 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-certificates\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469734 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-ca-trust-extracted\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469753 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-installation-pull-secrets\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469788 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnwp4\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-kube-api-access-xnwp4\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.469812 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-tls\") pod \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\" (UID: \"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26\") " Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.471155 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.471390 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.475110 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.475627 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.477841 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-kube-api-access-xnwp4" (OuterVolumeSpecName: "kube-api-access-xnwp4") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "kube-api-access-xnwp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.478782 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.486337 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.501207 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" (UID: "cc8ed586-cb64-4521-a4ed-e2dfa1a39f26"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571907 4679 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571948 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571957 4679 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571968 4679 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571977 4679 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571986 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnwp4\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-kube-api-access-xnwp4\") on node \"crc\" DevicePath \"\"" Dec 02 10:25:59 crc kubenswrapper[4679]: I1202 10:25:59.571995 4679 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:26:00 crc kubenswrapper[4679]: I1202 10:26:00.285109 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" event={"ID":"cc8ed586-cb64-4521-a4ed-e2dfa1a39f26","Type":"ContainerDied","Data":"d1d6721be96aa935f23fee51dd81fa8a1bbf9eed3883ac7b1491235abb5d556d"} Dec 02 10:26:00 crc kubenswrapper[4679]: I1202 10:26:00.285170 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h2ljp" Dec 02 10:26:00 crc kubenswrapper[4679]: I1202 10:26:00.286120 4679 scope.go:117] "RemoveContainer" containerID="7bb13bdcf48b735fe12f95e6ba5a297afa05925396a7c2f6b96ddb17d137e050" Dec 02 10:26:00 crc kubenswrapper[4679]: I1202 10:26:00.317552 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h2ljp"] Dec 02 10:26:00 crc kubenswrapper[4679]: I1202 10:26:00.322045 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h2ljp"] Dec 02 10:26:00 crc kubenswrapper[4679]: I1202 10:26:00.920520 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" path="/var/lib/kubelet/pods/cc8ed586-cb64-4521-a4ed-e2dfa1a39f26/volumes" Dec 02 10:26:16 crc kubenswrapper[4679]: I1202 10:26:16.931376 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:26:16 crc kubenswrapper[4679]: I1202 10:26:16.932116 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:26:16 crc kubenswrapper[4679]: I1202 10:26:16.932174 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:26:16 crc kubenswrapper[4679]: I1202 10:26:16.932897 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82a102b5f7c254ea6255ad947fef5a434fb11ff49e4dbe4618063250f135f38a"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:26:16 crc kubenswrapper[4679]: I1202 10:26:16.932994 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://82a102b5f7c254ea6255ad947fef5a434fb11ff49e4dbe4618063250f135f38a" gracePeriod=600 Dec 02 10:26:17 crc kubenswrapper[4679]: I1202 10:26:17.396542 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="82a102b5f7c254ea6255ad947fef5a434fb11ff49e4dbe4618063250f135f38a" exitCode=0 Dec 02 10:26:17 crc kubenswrapper[4679]: I1202 10:26:17.396584 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"82a102b5f7c254ea6255ad947fef5a434fb11ff49e4dbe4618063250f135f38a"} Dec 02 10:26:17 crc kubenswrapper[4679]: I1202 10:26:17.396869 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"366ff952d0a78699e9a3954a944a0a1122afa8431e83a016c9f4d6d35d914975"} Dec 02 10:26:17 crc kubenswrapper[4679]: I1202 10:26:17.396886 4679 scope.go:117] "RemoveContainer" containerID="264a3b1f599c8989075623acd8851f3a6226a47594e9a635be85a24102593311" Dec 02 10:28:07 crc kubenswrapper[4679]: I1202 10:28:07.107129 4679 scope.go:117] "RemoveContainer" containerID="66ee60dfb30afe85c5fa4ef583569cbb15b12c26c6c6a8f2b434b71e56f65847" Dec 02 10:28:46 crc kubenswrapper[4679]: I1202 10:28:46.931634 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:28:46 crc kubenswrapper[4679]: I1202 10:28:46.932341 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:29:07 crc kubenswrapper[4679]: I1202 10:29:07.148882 4679 scope.go:117] "RemoveContainer" containerID="69c6bfab086d5136a7ef6ce7b5f85730936be6d8960dd26d3436953a479e822c" Dec 02 10:29:07 crc kubenswrapper[4679]: I1202 10:29:07.168244 4679 scope.go:117] "RemoveContainer" containerID="a3b8b36118f67bd6af2327e1ab657d5c9902c0c0b79bca5f7fdf1307e463ab00" Dec 02 10:29:16 crc kubenswrapper[4679]: I1202 10:29:16.931831 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:29:16 crc kubenswrapper[4679]: I1202 10:29:16.932606 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:29:46 crc kubenswrapper[4679]: I1202 10:29:46.931789 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:29:46 crc kubenswrapper[4679]: I1202 10:29:46.933414 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:29:46 crc kubenswrapper[4679]: I1202 10:29:46.933480 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:29:46 crc kubenswrapper[4679]: I1202 10:29:46.934091 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"366ff952d0a78699e9a3954a944a0a1122afa8431e83a016c9f4d6d35d914975"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:29:46 crc kubenswrapper[4679]: I1202 10:29:46.934171 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://366ff952d0a78699e9a3954a944a0a1122afa8431e83a016c9f4d6d35d914975" gracePeriod=600 Dec 02 10:29:47 crc kubenswrapper[4679]: I1202 10:29:47.647780 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="366ff952d0a78699e9a3954a944a0a1122afa8431e83a016c9f4d6d35d914975" exitCode=0 Dec 02 10:29:47 crc kubenswrapper[4679]: I1202 10:29:47.647816 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"366ff952d0a78699e9a3954a944a0a1122afa8431e83a016c9f4d6d35d914975"} Dec 02 10:29:47 crc kubenswrapper[4679]: I1202 10:29:47.648164 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"09c15e7820d4c5f93b890b6f7afc7d06c4c33bed2214883f94f5eae437e987f2"} Dec 02 10:29:47 crc kubenswrapper[4679]: I1202 10:29:47.648200 4679 scope.go:117] "RemoveContainer" containerID="82a102b5f7c254ea6255ad947fef5a434fb11ff49e4dbe4618063250f135f38a" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.173446 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n"] Dec 02 10:30:00 crc kubenswrapper[4679]: E1202 10:30:00.174339 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" containerName="registry" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.174355 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" containerName="registry" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.174474 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8ed586-cb64-4521-a4ed-e2dfa1a39f26" containerName="registry" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.174925 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.177349 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.177363 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.182481 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n"] Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.244425 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdvwd\" (UniqueName: \"kubernetes.io/projected/18bd6ce5-4362-4262-99f9-cef7c2f0d104-kube-api-access-vdvwd\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.244504 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18bd6ce5-4362-4262-99f9-cef7c2f0d104-config-volume\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.244716 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18bd6ce5-4362-4262-99f9-cef7c2f0d104-secret-volume\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.346508 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdvwd\" (UniqueName: \"kubernetes.io/projected/18bd6ce5-4362-4262-99f9-cef7c2f0d104-kube-api-access-vdvwd\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.346583 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18bd6ce5-4362-4262-99f9-cef7c2f0d104-config-volume\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.346640 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18bd6ce5-4362-4262-99f9-cef7c2f0d104-secret-volume\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.348502 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18bd6ce5-4362-4262-99f9-cef7c2f0d104-config-volume\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.356815 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18bd6ce5-4362-4262-99f9-cef7c2f0d104-secret-volume\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.365816 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdvwd\" (UniqueName: \"kubernetes.io/projected/18bd6ce5-4362-4262-99f9-cef7c2f0d104-kube-api-access-vdvwd\") pod \"collect-profiles-29411190-l245n\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.494863 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.699458 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n"] Dec 02 10:30:00 crc kubenswrapper[4679]: I1202 10:30:00.724330 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" event={"ID":"18bd6ce5-4362-4262-99f9-cef7c2f0d104","Type":"ContainerStarted","Data":"6f4819a5d6aa4dccc63304a23a1c6c5623cad521609a148c563b25f86aeaf205"} Dec 02 10:30:01 crc kubenswrapper[4679]: I1202 10:30:01.730079 4679 generic.go:334] "Generic (PLEG): container finished" podID="18bd6ce5-4362-4262-99f9-cef7c2f0d104" containerID="b4a04b64c9e2ac8bd7aa4240181851555fdb3925ca820823fe1d57a6e9de3945" exitCode=0 Dec 02 10:30:01 crc kubenswrapper[4679]: I1202 10:30:01.730153 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" event={"ID":"18bd6ce5-4362-4262-99f9-cef7c2f0d104","Type":"ContainerDied","Data":"b4a04b64c9e2ac8bd7aa4240181851555fdb3925ca820823fe1d57a6e9de3945"} Dec 02 10:30:02 crc kubenswrapper[4679]: I1202 10:30:02.991861 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.081188 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdvwd\" (UniqueName: \"kubernetes.io/projected/18bd6ce5-4362-4262-99f9-cef7c2f0d104-kube-api-access-vdvwd\") pod \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.081246 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18bd6ce5-4362-4262-99f9-cef7c2f0d104-config-volume\") pod \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.081326 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18bd6ce5-4362-4262-99f9-cef7c2f0d104-secret-volume\") pod \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\" (UID: \"18bd6ce5-4362-4262-99f9-cef7c2f0d104\") " Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.081981 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18bd6ce5-4362-4262-99f9-cef7c2f0d104-config-volume" (OuterVolumeSpecName: "config-volume") pod "18bd6ce5-4362-4262-99f9-cef7c2f0d104" (UID: "18bd6ce5-4362-4262-99f9-cef7c2f0d104"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.086208 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18bd6ce5-4362-4262-99f9-cef7c2f0d104-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "18bd6ce5-4362-4262-99f9-cef7c2f0d104" (UID: "18bd6ce5-4362-4262-99f9-cef7c2f0d104"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.087075 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18bd6ce5-4362-4262-99f9-cef7c2f0d104-kube-api-access-vdvwd" (OuterVolumeSpecName: "kube-api-access-vdvwd") pod "18bd6ce5-4362-4262-99f9-cef7c2f0d104" (UID: "18bd6ce5-4362-4262-99f9-cef7c2f0d104"). InnerVolumeSpecName "kube-api-access-vdvwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.183244 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdvwd\" (UniqueName: \"kubernetes.io/projected/18bd6ce5-4362-4262-99f9-cef7c2f0d104-kube-api-access-vdvwd\") on node \"crc\" DevicePath \"\"" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.183279 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18bd6ce5-4362-4262-99f9-cef7c2f0d104-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.183288 4679 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/18bd6ce5-4362-4262-99f9-cef7c2f0d104-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.746862 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" event={"ID":"18bd6ce5-4362-4262-99f9-cef7c2f0d104","Type":"ContainerDied","Data":"6f4819a5d6aa4dccc63304a23a1c6c5623cad521609a148c563b25f86aeaf205"} Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.746906 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f4819a5d6aa4dccc63304a23a1c6c5623cad521609a148c563b25f86aeaf205" Dec 02 10:30:03 crc kubenswrapper[4679]: I1202 10:30:03.747019 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.174074 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dhkj7"] Dec 02 10:30:55 crc kubenswrapper[4679]: E1202 10:30:55.175011 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18bd6ce5-4362-4262-99f9-cef7c2f0d104" containerName="collect-profiles" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.175030 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="18bd6ce5-4362-4262-99f9-cef7c2f0d104" containerName="collect-profiles" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.175162 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="18bd6ce5-4362-4262-99f9-cef7c2f0d104" containerName="collect-profiles" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.175612 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.180124 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.180205 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.184210 4679 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-79q5c" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.189222 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dhkj7"] Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.196528 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-rnwm9"] Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.197164 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-rnwm9" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.200557 4679 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-fzdlx" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.207018 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-lhvkg"] Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.207936 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.216925 4679 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-x9n6f" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.225265 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-lhvkg"] Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.231566 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-rnwm9"] Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.345945 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltgvz\" (UniqueName: \"kubernetes.io/projected/1da80a36-7d00-4055-ba3e-1e84a58e2748-kube-api-access-ltgvz\") pod \"cert-manager-5b446d88c5-rnwm9\" (UID: \"1da80a36-7d00-4055-ba3e-1e84a58e2748\") " pod="cert-manager/cert-manager-5b446d88c5-rnwm9" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.345989 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhgrl\" (UniqueName: \"kubernetes.io/projected/227ec57c-bcbe-4725-8726-3411f96b09ac-kube-api-access-qhgrl\") pod \"cert-manager-cainjector-7f985d654d-dhkj7\" (UID: \"227ec57c-bcbe-4725-8726-3411f96b09ac\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.346019 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvsmr\" (UniqueName: \"kubernetes.io/projected/f91be00f-75e8-4a8b-aacc-d881b6af57a8-kube-api-access-gvsmr\") pod \"cert-manager-webhook-5655c58dd6-lhvkg\" (UID: \"f91be00f-75e8-4a8b-aacc-d881b6af57a8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.447911 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltgvz\" (UniqueName: \"kubernetes.io/projected/1da80a36-7d00-4055-ba3e-1e84a58e2748-kube-api-access-ltgvz\") pod \"cert-manager-5b446d88c5-rnwm9\" (UID: \"1da80a36-7d00-4055-ba3e-1e84a58e2748\") " pod="cert-manager/cert-manager-5b446d88c5-rnwm9" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.447980 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhgrl\" (UniqueName: \"kubernetes.io/projected/227ec57c-bcbe-4725-8726-3411f96b09ac-kube-api-access-qhgrl\") pod \"cert-manager-cainjector-7f985d654d-dhkj7\" (UID: \"227ec57c-bcbe-4725-8726-3411f96b09ac\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.448024 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvsmr\" (UniqueName: \"kubernetes.io/projected/f91be00f-75e8-4a8b-aacc-d881b6af57a8-kube-api-access-gvsmr\") pod \"cert-manager-webhook-5655c58dd6-lhvkg\" (UID: \"f91be00f-75e8-4a8b-aacc-d881b6af57a8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.467689 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltgvz\" (UniqueName: \"kubernetes.io/projected/1da80a36-7d00-4055-ba3e-1e84a58e2748-kube-api-access-ltgvz\") pod \"cert-manager-5b446d88c5-rnwm9\" (UID: \"1da80a36-7d00-4055-ba3e-1e84a58e2748\") " pod="cert-manager/cert-manager-5b446d88c5-rnwm9" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.467908 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhgrl\" (UniqueName: \"kubernetes.io/projected/227ec57c-bcbe-4725-8726-3411f96b09ac-kube-api-access-qhgrl\") pod \"cert-manager-cainjector-7f985d654d-dhkj7\" (UID: \"227ec57c-bcbe-4725-8726-3411f96b09ac\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.469599 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvsmr\" (UniqueName: \"kubernetes.io/projected/f91be00f-75e8-4a8b-aacc-d881b6af57a8-kube-api-access-gvsmr\") pod \"cert-manager-webhook-5655c58dd6-lhvkg\" (UID: \"f91be00f-75e8-4a8b-aacc-d881b6af57a8\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.496350 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.512096 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-rnwm9" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.524060 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.948583 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dhkj7"] Dec 02 10:30:55 crc kubenswrapper[4679]: I1202 10:30:55.956731 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:30:56 crc kubenswrapper[4679]: I1202 10:30:56.006539 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-rnwm9"] Dec 02 10:30:56 crc kubenswrapper[4679]: I1202 10:30:56.011550 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-lhvkg"] Dec 02 10:30:56 crc kubenswrapper[4679]: W1202 10:30:56.027029 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf91be00f_75e8_4a8b_aacc_d881b6af57a8.slice/crio-6a7ef90c218fe1f3b02e9a209b0aa750a8a1b9355ce0ae002c3acbde8844a83e WatchSource:0}: Error finding container 6a7ef90c218fe1f3b02e9a209b0aa750a8a1b9355ce0ae002c3acbde8844a83e: Status 404 returned error can't find the container with id 6a7ef90c218fe1f3b02e9a209b0aa750a8a1b9355ce0ae002c3acbde8844a83e Dec 02 10:30:56 crc kubenswrapper[4679]: I1202 10:30:56.061985 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" event={"ID":"227ec57c-bcbe-4725-8726-3411f96b09ac","Type":"ContainerStarted","Data":"31583cc9bc325886d62f016612f9b72b319a1484178fb67634bc0a0bd01412e4"} Dec 02 10:30:56 crc kubenswrapper[4679]: I1202 10:30:56.062871 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" event={"ID":"f91be00f-75e8-4a8b-aacc-d881b6af57a8","Type":"ContainerStarted","Data":"6a7ef90c218fe1f3b02e9a209b0aa750a8a1b9355ce0ae002c3acbde8844a83e"} Dec 02 10:30:56 crc kubenswrapper[4679]: I1202 10:30:56.063752 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-rnwm9" event={"ID":"1da80a36-7d00-4055-ba3e-1e84a58e2748","Type":"ContainerStarted","Data":"de460abf8b9dba6b01c8c23ba7ed39ca82b6d559c556d007acd01371f92b5e10"} Dec 02 10:30:59 crc kubenswrapper[4679]: I1202 10:30:59.095894 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" event={"ID":"227ec57c-bcbe-4725-8726-3411f96b09ac","Type":"ContainerStarted","Data":"03b7481d114a74aa03e8f682bf94dfa916880f09c6b9bb88c12ba70757ce083f"} Dec 02 10:30:59 crc kubenswrapper[4679]: I1202 10:30:59.100619 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-rnwm9" event={"ID":"1da80a36-7d00-4055-ba3e-1e84a58e2748","Type":"ContainerStarted","Data":"2e43a8a3b25f2b568343443ccc212041daaf10d879d0e77e695620e8034ceb6e"} Dec 02 10:30:59 crc kubenswrapper[4679]: I1202 10:30:59.111792 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-dhkj7" podStartSLOduration=1.325310706 podStartE2EDuration="4.11176462s" podCreationTimestamp="2025-12-02 10:30:55 +0000 UTC" firstStartedPulling="2025-12-02 10:30:55.956479567 +0000 UTC m=+709.286618437" lastFinishedPulling="2025-12-02 10:30:58.742933491 +0000 UTC m=+712.073072351" observedRunningTime="2025-12-02 10:30:59.111612785 +0000 UTC m=+712.441751645" watchObservedRunningTime="2025-12-02 10:30:59.11176462 +0000 UTC m=+712.441903480" Dec 02 10:30:59 crc kubenswrapper[4679]: I1202 10:30:59.137811 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-rnwm9" podStartSLOduration=1.476960831 podStartE2EDuration="4.137783197s" podCreationTimestamp="2025-12-02 10:30:55 +0000 UTC" firstStartedPulling="2025-12-02 10:30:56.018496787 +0000 UTC m=+709.348635657" lastFinishedPulling="2025-12-02 10:30:58.679319173 +0000 UTC m=+712.009458023" observedRunningTime="2025-12-02 10:30:59.131444082 +0000 UTC m=+712.461582942" watchObservedRunningTime="2025-12-02 10:30:59.137783197 +0000 UTC m=+712.467922057" Dec 02 10:31:00 crc kubenswrapper[4679]: I1202 10:31:00.108073 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" event={"ID":"f91be00f-75e8-4a8b-aacc-d881b6af57a8","Type":"ContainerStarted","Data":"61ed67347ee1208fe5929def3453056503d402745cae1087c4c4008079e4069f"} Dec 02 10:31:00 crc kubenswrapper[4679]: I1202 10:31:00.128016 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" podStartSLOduration=1.399565749 podStartE2EDuration="5.127988228s" podCreationTimestamp="2025-12-02 10:30:55 +0000 UTC" firstStartedPulling="2025-12-02 10:30:56.03101508 +0000 UTC m=+709.361153940" lastFinishedPulling="2025-12-02 10:30:59.759437559 +0000 UTC m=+713.089576419" observedRunningTime="2025-12-02 10:31:00.125495912 +0000 UTC m=+713.455634782" watchObservedRunningTime="2025-12-02 10:31:00.127988228 +0000 UTC m=+713.458127098" Dec 02 10:31:00 crc kubenswrapper[4679]: I1202 10:31:00.524588 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.527287 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-lhvkg" Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.667404 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pjb6d"] Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.667957 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="sbdb" containerID="cri-o://97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.668014 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.668082 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-node" containerID="cri-o://5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.668116 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-acl-logging" containerID="cri-o://818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.668186 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="northd" containerID="cri-o://7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.668181 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="nbdb" containerID="cri-o://e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.667943 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-controller" containerID="cri-o://014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.712034 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" containerID="cri-o://5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" gracePeriod=30 Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.959316 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/3.log" Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.961590 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovn-acl-logging/0.log" Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.962141 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovn-controller/0.log" Dec 02 10:31:05 crc kubenswrapper[4679]: I1202 10:31:05.962515 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032001 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4b5pn"] Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032238 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032254 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032267 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032280 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032295 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kubecfg-setup" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032329 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kubecfg-setup" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032341 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="northd" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032349 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="northd" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032363 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-node" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032372 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-node" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032382 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="nbdb" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032391 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="nbdb" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032431 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032441 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032453 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-acl-logging" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032463 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-acl-logging" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032476 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032488 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032500 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032509 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032524 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032534 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032550 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="sbdb" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032560 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="sbdb" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032701 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032715 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="kube-rbac-proxy-node" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032731 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032744 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="northd" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032758 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032772 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032792 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="nbdb" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032805 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032820 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovn-acl-logging" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032837 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="sbdb" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.032977 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.032987 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.033115 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.033135 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerName="ovnkube-controller" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.035850 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078630 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-env-overrides\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078687 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-netd\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078741 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-systemd\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078765 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-ovn-kubernetes\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078818 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078855 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-netns\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078900 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-node-log\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078922 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-systemd-units\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078972 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-kubelet\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.078998 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovn-node-metrics-cert\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079045 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-var-lib-openvswitch\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079079 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-etc-openvswitch\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079130 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-openvswitch\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079171 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-bin\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079223 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-ovn\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079247 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-config\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079127 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079160 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-node-log" (OuterVolumeSpecName: "node-log") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079177 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079581 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079581 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079613 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079632 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079645 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079658 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079671 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079509 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmbpj\" (UniqueName: \"kubernetes.io/projected/a66c9e2c-2ca3-4348-84cc-19f365505c9e-kube-api-access-cmbpj\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079733 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-log-socket\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079766 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-slash\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079788 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-script-lib\") pod \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\" (UID: \"a66c9e2c-2ca3-4348-84cc-19f365505c9e\") " Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079680 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079696 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079918 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.079956 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-log-socket" (OuterVolumeSpecName: "log-socket") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080168 4679 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080178 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-slash" (OuterVolumeSpecName: "host-slash") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080220 4679 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080234 4679 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080247 4679 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080258 4679 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080269 4679 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-node-log\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080278 4679 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080293 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080287 4679 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080460 4679 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080474 4679 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080484 4679 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080483 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080494 4679 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.080528 4679 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.084490 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a66c9e2c-2ca3-4348-84cc-19f365505c9e-kube-api-access-cmbpj" (OuterVolumeSpecName: "kube-api-access-cmbpj") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "kube-api-access-cmbpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.084497 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.091332 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a66c9e2c-2ca3-4348-84cc-19f365505c9e" (UID: "a66c9e2c-2ca3-4348-84cc-19f365505c9e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.141432 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovnkube-controller/3.log" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.143480 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovn-acl-logging/0.log" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.143981 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pjb6d_a66c9e2c-2ca3-4348-84cc-19f365505c9e/ovn-controller/0.log" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144409 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" exitCode=0 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144480 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144533 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144568 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144618 4679 scope.go:117] "RemoveContainer" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144507 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" exitCode=0 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144754 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" exitCode=0 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144825 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" exitCode=0 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144900 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" exitCode=0 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144968 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" exitCode=0 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145081 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" exitCode=143 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145151 4679 generic.go:334] "Generic (PLEG): container finished" podID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" exitCode=143 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.144800 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145369 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145467 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145545 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145610 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145670 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145716 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145772 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145819 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145863 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145911 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.145989 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146054 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146148 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146278 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146419 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146625 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146682 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146739 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146786 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146874 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.146981 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147037 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147094 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147152 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147204 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147251 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147321 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147368 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147436 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147484 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147527 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147578 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147646 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147713 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147779 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pjb6d" event={"ID":"a66c9e2c-2ca3-4348-84cc-19f365505c9e","Type":"ContainerDied","Data":"01cb27479623951c56ca929f70f828f97b7d6423b7afa1e9ad4da9d63a886485"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147833 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147886 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147932 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147982 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.148034 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.148086 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.148137 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.148189 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.148238 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.148287 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.147521 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/2.log" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.149353 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/1.log" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.149409 4679 generic.go:334] "Generic (PLEG): container finished" podID="06f77afa-f710-437b-9cfe-10959bdc1ac8" containerID="37cca3a05037016399c4df02f078b7a1a8d43fa20e09c359c1b34e975cdc82f7" exitCode=2 Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.149441 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerDied","Data":"37cca3a05037016399c4df02f078b7a1a8d43fa20e09c359c1b34e975cdc82f7"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.149464 4679 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab"} Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.149996 4679 scope.go:117] "RemoveContainer" containerID="37cca3a05037016399c4df02f078b7a1a8d43fa20e09c359c1b34e975cdc82f7" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.150289 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-r5sft_openshift-multus(06f77afa-f710-437b-9cfe-10959bdc1ac8)\"" pod="openshift-multus/multus-r5sft" podUID="06f77afa-f710-437b-9cfe-10959bdc1ac8" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.175375 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184755 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-cni-bin\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184812 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-env-overrides\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184838 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-kubelet\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184904 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-ovn\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184922 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-run-netns\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184962 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-var-lib-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184978 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-cni-netd\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.184993 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-log-socket\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185010 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-slash\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185043 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-run-ovn-kubernetes\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185062 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185080 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-node-log\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185121 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-systemd-units\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185140 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-etc-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185154 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185198 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-ovnkube-config\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185216 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-ovnkube-script-lib\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185234 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1fd09944-1660-4a68-b122-121114f7770f-ovn-node-metrics-cert\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185268 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-systemd\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185286 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgftk\" (UniqueName: \"kubernetes.io/projected/1fd09944-1660-4a68-b122-121114f7770f-kube-api-access-dgftk\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185344 4679 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185355 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmbpj\" (UniqueName: \"kubernetes.io/projected/a66c9e2c-2ca3-4348-84cc-19f365505c9e-kube-api-access-cmbpj\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185364 4679 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185372 4679 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-log-socket\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185382 4679 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-host-slash\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185408 4679 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a66c9e2c-2ca3-4348-84cc-19f365505c9e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185417 4679 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a66c9e2c-2ca3-4348-84cc-19f365505c9e-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.185956 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pjb6d"] Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.192910 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pjb6d"] Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.209520 4679 scope.go:117] "RemoveContainer" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.225415 4679 scope.go:117] "RemoveContainer" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.238082 4679 scope.go:117] "RemoveContainer" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.254679 4679 scope.go:117] "RemoveContainer" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.269148 4679 scope.go:117] "RemoveContainer" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.282524 4679 scope.go:117] "RemoveContainer" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286217 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-ovnkube-script-lib\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286250 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1fd09944-1660-4a68-b122-121114f7770f-ovn-node-metrics-cert\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286282 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-systemd\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286319 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgftk\" (UniqueName: \"kubernetes.io/projected/1fd09944-1660-4a68-b122-121114f7770f-kube-api-access-dgftk\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286336 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-cni-bin\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286349 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-kubelet\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286369 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-env-overrides\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286403 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-ovn\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286421 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-run-netns\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286439 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-var-lib-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286471 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-cni-netd\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286488 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-log-socket\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286503 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-slash\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286521 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-run-ovn-kubernetes\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286545 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286563 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-node-log\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286593 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-systemd-units\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286613 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-etc-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286626 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286649 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-ovnkube-config\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.286982 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-run-netns\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287011 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-run-ovn-kubernetes\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287068 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-systemd-units\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287126 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-ovn\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287215 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-var-lib-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287322 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287019 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-cni-bin\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287277 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-etc-openvswitch\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287246 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-ovnkube-config\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287538 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-node-log\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287520 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287523 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-cni-netd\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287536 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-env-overrides\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287557 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-slash\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287582 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-host-kubelet\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287563 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-run-systemd\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.287501 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1fd09944-1660-4a68-b122-121114f7770f-log-socket\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.288046 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1fd09944-1660-4a68-b122-121114f7770f-ovnkube-script-lib\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.290516 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1fd09944-1660-4a68-b122-121114f7770f-ovn-node-metrics-cert\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.295461 4679 scope.go:117] "RemoveContainer" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.304823 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgftk\" (UniqueName: \"kubernetes.io/projected/1fd09944-1660-4a68-b122-121114f7770f-kube-api-access-dgftk\") pod \"ovnkube-node-4b5pn\" (UID: \"1fd09944-1660-4a68-b122-121114f7770f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.308606 4679 scope.go:117] "RemoveContainer" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.321977 4679 scope.go:117] "RemoveContainer" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.322516 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": container with ID starting with 5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131 not found: ID does not exist" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.322541 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} err="failed to get container status \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": rpc error: code = NotFound desc = could not find container \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": container with ID starting with 5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.322561 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.322866 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": container with ID starting with 01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c not found: ID does not exist" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.322885 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} err="failed to get container status \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": rpc error: code = NotFound desc = could not find container \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": container with ID starting with 01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.322896 4679 scope.go:117] "RemoveContainer" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.323096 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": container with ID starting with 97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85 not found: ID does not exist" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.323115 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} err="failed to get container status \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": rpc error: code = NotFound desc = could not find container \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": container with ID starting with 97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.323126 4679 scope.go:117] "RemoveContainer" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.323400 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": container with ID starting with e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7 not found: ID does not exist" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.323422 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} err="failed to get container status \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": rpc error: code = NotFound desc = could not find container \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": container with ID starting with e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.323434 4679 scope.go:117] "RemoveContainer" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.324625 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": container with ID starting with 7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76 not found: ID does not exist" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.324645 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} err="failed to get container status \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": rpc error: code = NotFound desc = could not find container \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": container with ID starting with 7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.324657 4679 scope.go:117] "RemoveContainer" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.324891 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": container with ID starting with aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7 not found: ID does not exist" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.324954 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} err="failed to get container status \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": rpc error: code = NotFound desc = could not find container \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": container with ID starting with aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.324984 4679 scope.go:117] "RemoveContainer" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.325339 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": container with ID starting with 5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109 not found: ID does not exist" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.325359 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} err="failed to get container status \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": rpc error: code = NotFound desc = could not find container \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": container with ID starting with 5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.325375 4679 scope.go:117] "RemoveContainer" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.325618 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": container with ID starting with 818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721 not found: ID does not exist" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.325638 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} err="failed to get container status \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": rpc error: code = NotFound desc = could not find container \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": container with ID starting with 818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.325653 4679 scope.go:117] "RemoveContainer" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.325944 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": container with ID starting with 014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d not found: ID does not exist" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.325965 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} err="failed to get container status \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": rpc error: code = NotFound desc = could not find container \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": container with ID starting with 014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.325980 4679 scope.go:117] "RemoveContainer" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" Dec 02 10:31:06 crc kubenswrapper[4679]: E1202 10:31:06.326200 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": container with ID starting with 666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d not found: ID does not exist" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326216 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} err="failed to get container status \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": rpc error: code = NotFound desc = could not find container \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": container with ID starting with 666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326228 4679 scope.go:117] "RemoveContainer" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326437 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} err="failed to get container status \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": rpc error: code = NotFound desc = could not find container \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": container with ID starting with 5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326463 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326670 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} err="failed to get container status \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": rpc error: code = NotFound desc = could not find container \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": container with ID starting with 01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326692 4679 scope.go:117] "RemoveContainer" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326943 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} err="failed to get container status \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": rpc error: code = NotFound desc = could not find container \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": container with ID starting with 97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.326991 4679 scope.go:117] "RemoveContainer" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327229 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} err="failed to get container status \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": rpc error: code = NotFound desc = could not find container \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": container with ID starting with e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327251 4679 scope.go:117] "RemoveContainer" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327440 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} err="failed to get container status \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": rpc error: code = NotFound desc = could not find container \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": container with ID starting with 7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327458 4679 scope.go:117] "RemoveContainer" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327628 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} err="failed to get container status \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": rpc error: code = NotFound desc = could not find container \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": container with ID starting with aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327643 4679 scope.go:117] "RemoveContainer" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327831 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} err="failed to get container status \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": rpc error: code = NotFound desc = could not find container \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": container with ID starting with 5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.327847 4679 scope.go:117] "RemoveContainer" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328016 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} err="failed to get container status \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": rpc error: code = NotFound desc = could not find container \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": container with ID starting with 818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328043 4679 scope.go:117] "RemoveContainer" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328326 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} err="failed to get container status \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": rpc error: code = NotFound desc = could not find container \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": container with ID starting with 014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328343 4679 scope.go:117] "RemoveContainer" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328631 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} err="failed to get container status \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": rpc error: code = NotFound desc = could not find container \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": container with ID starting with 666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328651 4679 scope.go:117] "RemoveContainer" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328886 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} err="failed to get container status \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": rpc error: code = NotFound desc = could not find container \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": container with ID starting with 5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.328903 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329099 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} err="failed to get container status \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": rpc error: code = NotFound desc = could not find container \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": container with ID starting with 01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329136 4679 scope.go:117] "RemoveContainer" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329355 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} err="failed to get container status \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": rpc error: code = NotFound desc = could not find container \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": container with ID starting with 97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329370 4679 scope.go:117] "RemoveContainer" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329546 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} err="failed to get container status \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": rpc error: code = NotFound desc = could not find container \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": container with ID starting with e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329561 4679 scope.go:117] "RemoveContainer" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329721 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} err="failed to get container status \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": rpc error: code = NotFound desc = could not find container \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": container with ID starting with 7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329745 4679 scope.go:117] "RemoveContainer" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329932 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} err="failed to get container status \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": rpc error: code = NotFound desc = could not find container \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": container with ID starting with aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.329956 4679 scope.go:117] "RemoveContainer" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330136 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} err="failed to get container status \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": rpc error: code = NotFound desc = could not find container \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": container with ID starting with 5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330157 4679 scope.go:117] "RemoveContainer" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330424 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} err="failed to get container status \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": rpc error: code = NotFound desc = could not find container \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": container with ID starting with 818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330442 4679 scope.go:117] "RemoveContainer" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330657 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} err="failed to get container status \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": rpc error: code = NotFound desc = could not find container \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": container with ID starting with 014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330689 4679 scope.go:117] "RemoveContainer" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330964 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} err="failed to get container status \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": rpc error: code = NotFound desc = could not find container \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": container with ID starting with 666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.330982 4679 scope.go:117] "RemoveContainer" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331242 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} err="failed to get container status \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": rpc error: code = NotFound desc = could not find container \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": container with ID starting with 5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331259 4679 scope.go:117] "RemoveContainer" containerID="01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331476 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c"} err="failed to get container status \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": rpc error: code = NotFound desc = could not find container \"01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c\": container with ID starting with 01588aba09998b63496f44420c96ff4a64700225047260f3987425388febe24c not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331496 4679 scope.go:117] "RemoveContainer" containerID="97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331747 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85"} err="failed to get container status \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": rpc error: code = NotFound desc = could not find container \"97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85\": container with ID starting with 97737cf834aa4257838934b2280db35416363af71493b4a9e297981c88db4c85 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331767 4679 scope.go:117] "RemoveContainer" containerID="e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331975 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7"} err="failed to get container status \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": rpc error: code = NotFound desc = could not find container \"e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7\": container with ID starting with e052a2e7dfc93f8b0a4f8849e1be900f6ca809565bc44d9da898a48bf2344dd7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.331991 4679 scope.go:117] "RemoveContainer" containerID="7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332168 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76"} err="failed to get container status \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": rpc error: code = NotFound desc = could not find container \"7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76\": container with ID starting with 7d0588a12eb94d52fd662d8f003d1a58792a431fb6ac2062b1f10246ed7e2a76 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332184 4679 scope.go:117] "RemoveContainer" containerID="aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332400 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7"} err="failed to get container status \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": rpc error: code = NotFound desc = could not find container \"aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7\": container with ID starting with aa9c1e5df0fb352cf54ad46e6b9999b2bd0a689f4aa1cb97ad48f29c544634d7 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332427 4679 scope.go:117] "RemoveContainer" containerID="5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332623 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109"} err="failed to get container status \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": rpc error: code = NotFound desc = could not find container \"5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109\": container with ID starting with 5394685dbf534782a5f7693ca7c36b39b0369b16cab8da5655f8850ccd476109 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332638 4679 scope.go:117] "RemoveContainer" containerID="818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332830 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721"} err="failed to get container status \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": rpc error: code = NotFound desc = could not find container \"818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721\": container with ID starting with 818c5cb52cd0723b8a4a9051e85268aac3c627b76a4c0cd5f51bf03f19a70721 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.332843 4679 scope.go:117] "RemoveContainer" containerID="014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.333134 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d"} err="failed to get container status \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": rpc error: code = NotFound desc = could not find container \"014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d\": container with ID starting with 014e90f53b0699963368540f345572f3b06964705e6916820da7b54100f5582d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.333215 4679 scope.go:117] "RemoveContainer" containerID="666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.333518 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d"} err="failed to get container status \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": rpc error: code = NotFound desc = could not find container \"666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d\": container with ID starting with 666ce1fa299f933970805b4cab4f060cbbf89cb2c1f1605624b393424649699d not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.333536 4679 scope.go:117] "RemoveContainer" containerID="5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.333775 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131"} err="failed to get container status \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": rpc error: code = NotFound desc = could not find container \"5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131\": container with ID starting with 5e0713281d4db580affc3abd1202c7b00e719d7e31e8d5d75c2f606c1fcf2131 not found: ID does not exist" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.353440 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:06 crc kubenswrapper[4679]: I1202 10:31:06.918659 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a66c9e2c-2ca3-4348-84cc-19f365505c9e" path="/var/lib/kubelet/pods/a66c9e2c-2ca3-4348-84cc-19f365505c9e/volumes" Dec 02 10:31:07 crc kubenswrapper[4679]: I1202 10:31:07.159408 4679 generic.go:334] "Generic (PLEG): container finished" podID="1fd09944-1660-4a68-b122-121114f7770f" containerID="78e71227ffa77c7df2fdaaeff5b19e83b15ee1304340eb98e78a578722afa3d3" exitCode=0 Dec 02 10:31:07 crc kubenswrapper[4679]: I1202 10:31:07.159535 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerDied","Data":"78e71227ffa77c7df2fdaaeff5b19e83b15ee1304340eb98e78a578722afa3d3"} Dec 02 10:31:07 crc kubenswrapper[4679]: I1202 10:31:07.159614 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"e05330eb400b86f7266344037f72e9ee25df385413f8556ca894aa27f49d5844"} Dec 02 10:31:07 crc kubenswrapper[4679]: I1202 10:31:07.213637 4679 scope.go:117] "RemoveContainer" containerID="97b5c818c178a20cbb9afce9bedb8e7473acae93329b96c637205812c5b0c6ab" Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.166449 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/2.log" Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.169202 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"8f7472b6315af3eead8b30786c2d1c49b103ac4f2c14d6e102ee27ceb7b6e816"} Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.169240 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"330262b2a34854e95df4a94c189ba5678903fd3c36cbd7c7777eae8e1e007cb6"} Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.169253 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"9612125aa09c650b8059ce5beacc9380ca63e6cea467306db2ec3131b96993d5"} Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.169267 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"74ec101535536b458c4c5990620880e3c71048461c664eaccd148799e11f623c"} Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.169316 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"4d81c724fcb66775a88302f09bd67f4ca0f27b81ba36eaaadf8247700be400d0"} Dec 02 10:31:08 crc kubenswrapper[4679]: I1202 10:31:08.169327 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"4e04a1c3df1097a7d54ffccb5712a4e60cc4936db1eb6a6fc1aced948bf24d26"} Dec 02 10:31:10 crc kubenswrapper[4679]: I1202 10:31:10.187261 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"2cb6b2213b772dd627de68b2b120be48e42c273758289d4f9d14a5531a59d54a"} Dec 02 10:31:13 crc kubenswrapper[4679]: I1202 10:31:13.205966 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" event={"ID":"1fd09944-1660-4a68-b122-121114f7770f","Type":"ContainerStarted","Data":"2e5c912efc59118787bac8b09cbdf6ff10e31d79d9062e0d60cf8294053290d7"} Dec 02 10:31:13 crc kubenswrapper[4679]: I1202 10:31:13.206605 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:13 crc kubenswrapper[4679]: I1202 10:31:13.206620 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:13 crc kubenswrapper[4679]: I1202 10:31:13.247197 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:13 crc kubenswrapper[4679]: I1202 10:31:13.270778 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" podStartSLOduration=7.270760351 podStartE2EDuration="7.270760351s" podCreationTimestamp="2025-12-02 10:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:31:13.245106648 +0000 UTC m=+726.575245558" watchObservedRunningTime="2025-12-02 10:31:13.270760351 +0000 UTC m=+726.600899211" Dec 02 10:31:14 crc kubenswrapper[4679]: I1202 10:31:14.212160 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:14 crc kubenswrapper[4679]: I1202 10:31:14.282227 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:17 crc kubenswrapper[4679]: I1202 10:31:17.909113 4679 scope.go:117] "RemoveContainer" containerID="37cca3a05037016399c4df02f078b7a1a8d43fa20e09c359c1b34e975cdc82f7" Dec 02 10:31:19 crc kubenswrapper[4679]: I1202 10:31:19.244411 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-r5sft_06f77afa-f710-437b-9cfe-10959bdc1ac8/kube-multus/2.log" Dec 02 10:31:19 crc kubenswrapper[4679]: I1202 10:31:19.245389 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-r5sft" event={"ID":"06f77afa-f710-437b-9cfe-10959bdc1ac8","Type":"ContainerStarted","Data":"6ec0c70d678283375c7818820b03e1b710b4d5a857f6739a426a140bb91c2c01"} Dec 02 10:31:36 crc kubenswrapper[4679]: I1202 10:31:36.379887 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4b5pn" Dec 02 10:31:46 crc kubenswrapper[4679]: I1202 10:31:46.879121 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86"] Dec 02 10:31:46 crc kubenswrapper[4679]: I1202 10:31:46.883448 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:46 crc kubenswrapper[4679]: I1202 10:31:46.887226 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 10:31:46 crc kubenswrapper[4679]: I1202 10:31:46.901004 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86"] Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.050832 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.051388 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfmsm\" (UniqueName: \"kubernetes.io/projected/ae8d2b3f-baec-4d47-be99-1d73bc31416d-kube-api-access-xfmsm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.051425 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.153090 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfmsm\" (UniqueName: \"kubernetes.io/projected/ae8d2b3f-baec-4d47-be99-1d73bc31416d-kube-api-access-xfmsm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.153157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.153214 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.154098 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.154404 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.184763 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfmsm\" (UniqueName: \"kubernetes.io/projected/ae8d2b3f-baec-4d47-be99-1d73bc31416d-kube-api-access-xfmsm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.260420 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:47 crc kubenswrapper[4679]: I1202 10:31:47.665910 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86"] Dec 02 10:31:48 crc kubenswrapper[4679]: I1202 10:31:48.412846 4679 generic.go:334] "Generic (PLEG): container finished" podID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerID="6bdfcd99257d9c5a64be14f682f815f38d5c319b5cf8071e39ac7add46e42cd1" exitCode=0 Dec 02 10:31:48 crc kubenswrapper[4679]: I1202 10:31:48.412908 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" event={"ID":"ae8d2b3f-baec-4d47-be99-1d73bc31416d","Type":"ContainerDied","Data":"6bdfcd99257d9c5a64be14f682f815f38d5c319b5cf8071e39ac7add46e42cd1"} Dec 02 10:31:48 crc kubenswrapper[4679]: I1202 10:31:48.414816 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" event={"ID":"ae8d2b3f-baec-4d47-be99-1d73bc31416d","Type":"ContainerStarted","Data":"296638cb5cd973cee90ebe5a443908778fa8344831c0d962a857fc81bb1b37bd"} Dec 02 10:31:48 crc kubenswrapper[4679]: I1202 10:31:48.836769 4679 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.136733 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-chrjz"] Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.137847 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.154283 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-chrjz"] Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.284630 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-catalog-content\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.284712 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtqqk\" (UniqueName: \"kubernetes.io/projected/a78093b3-ae68-4a47-a34e-57799e342169-kube-api-access-wtqqk\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.284814 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-utilities\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.386459 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtqqk\" (UniqueName: \"kubernetes.io/projected/a78093b3-ae68-4a47-a34e-57799e342169-kube-api-access-wtqqk\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.386598 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-utilities\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.386639 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-catalog-content\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.387227 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-utilities\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.387291 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-catalog-content\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.408711 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtqqk\" (UniqueName: \"kubernetes.io/projected/a78093b3-ae68-4a47-a34e-57799e342169-kube-api-access-wtqqk\") pod \"redhat-operators-chrjz\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.455710 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:49 crc kubenswrapper[4679]: I1202 10:31:49.884516 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-chrjz"] Dec 02 10:31:50 crc kubenswrapper[4679]: I1202 10:31:50.428058 4679 generic.go:334] "Generic (PLEG): container finished" podID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerID="4975fe7ae16b218e49b704fd3d44dd9c37f9204de045e1f2a2e38540b1504cb8" exitCode=0 Dec 02 10:31:50 crc kubenswrapper[4679]: I1202 10:31:50.428156 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" event={"ID":"ae8d2b3f-baec-4d47-be99-1d73bc31416d","Type":"ContainerDied","Data":"4975fe7ae16b218e49b704fd3d44dd9c37f9204de045e1f2a2e38540b1504cb8"} Dec 02 10:31:50 crc kubenswrapper[4679]: I1202 10:31:50.431451 4679 generic.go:334] "Generic (PLEG): container finished" podID="a78093b3-ae68-4a47-a34e-57799e342169" containerID="a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6" exitCode=0 Dec 02 10:31:50 crc kubenswrapper[4679]: I1202 10:31:50.431505 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chrjz" event={"ID":"a78093b3-ae68-4a47-a34e-57799e342169","Type":"ContainerDied","Data":"a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6"} Dec 02 10:31:50 crc kubenswrapper[4679]: I1202 10:31:50.431543 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chrjz" event={"ID":"a78093b3-ae68-4a47-a34e-57799e342169","Type":"ContainerStarted","Data":"e7f14f0eee13208d8ba73d9c4c1ddaba84832b4ec86e2f242873fd1db1a6888d"} Dec 02 10:31:51 crc kubenswrapper[4679]: I1202 10:31:51.442768 4679 generic.go:334] "Generic (PLEG): container finished" podID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerID="9142476c1e8c15f84dc4472a8ab87a0b581121dab4170bdf880469f994513582" exitCode=0 Dec 02 10:31:51 crc kubenswrapper[4679]: I1202 10:31:51.442830 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" event={"ID":"ae8d2b3f-baec-4d47-be99-1d73bc31416d","Type":"ContainerDied","Data":"9142476c1e8c15f84dc4472a8ab87a0b581121dab4170bdf880469f994513582"} Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.461995 4679 generic.go:334] "Generic (PLEG): container finished" podID="a78093b3-ae68-4a47-a34e-57799e342169" containerID="b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028" exitCode=0 Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.462174 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chrjz" event={"ID":"a78093b3-ae68-4a47-a34e-57799e342169","Type":"ContainerDied","Data":"b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028"} Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.721184 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.830145 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-util\") pod \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.830255 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfmsm\" (UniqueName: \"kubernetes.io/projected/ae8d2b3f-baec-4d47-be99-1d73bc31416d-kube-api-access-xfmsm\") pod \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.830375 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-bundle\") pod \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\" (UID: \"ae8d2b3f-baec-4d47-be99-1d73bc31416d\") " Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.830947 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-bundle" (OuterVolumeSpecName: "bundle") pod "ae8d2b3f-baec-4d47-be99-1d73bc31416d" (UID: "ae8d2b3f-baec-4d47-be99-1d73bc31416d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.831224 4679 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.842053 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8d2b3f-baec-4d47-be99-1d73bc31416d-kube-api-access-xfmsm" (OuterVolumeSpecName: "kube-api-access-xfmsm") pod "ae8d2b3f-baec-4d47-be99-1d73bc31416d" (UID: "ae8d2b3f-baec-4d47-be99-1d73bc31416d"). InnerVolumeSpecName "kube-api-access-xfmsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.848636 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-util" (OuterVolumeSpecName: "util") pod "ae8d2b3f-baec-4d47-be99-1d73bc31416d" (UID: "ae8d2b3f-baec-4d47-be99-1d73bc31416d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.933051 4679 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae8d2b3f-baec-4d47-be99-1d73bc31416d-util\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:52 crc kubenswrapper[4679]: I1202 10:31:52.933118 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfmsm\" (UniqueName: \"kubernetes.io/projected/ae8d2b3f-baec-4d47-be99-1d73bc31416d-kube-api-access-xfmsm\") on node \"crc\" DevicePath \"\"" Dec 02 10:31:53 crc kubenswrapper[4679]: I1202 10:31:53.470574 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" event={"ID":"ae8d2b3f-baec-4d47-be99-1d73bc31416d","Type":"ContainerDied","Data":"296638cb5cd973cee90ebe5a443908778fa8344831c0d962a857fc81bb1b37bd"} Dec 02 10:31:53 crc kubenswrapper[4679]: I1202 10:31:53.470923 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="296638cb5cd973cee90ebe5a443908778fa8344831c0d962a857fc81bb1b37bd" Dec 02 10:31:53 crc kubenswrapper[4679]: I1202 10:31:53.470641 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86" Dec 02 10:31:54 crc kubenswrapper[4679]: I1202 10:31:54.480121 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chrjz" event={"ID":"a78093b3-ae68-4a47-a34e-57799e342169","Type":"ContainerStarted","Data":"9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b"} Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.231376 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-chrjz" podStartSLOduration=5.173007371 podStartE2EDuration="8.231353633s" podCreationTimestamp="2025-12-02 10:31:49 +0000 UTC" firstStartedPulling="2025-12-02 10:31:50.43315495 +0000 UTC m=+763.763293810" lastFinishedPulling="2025-12-02 10:31:53.491501212 +0000 UTC m=+766.821640072" observedRunningTime="2025-12-02 10:31:54.507577204 +0000 UTC m=+767.837716094" watchObservedRunningTime="2025-12-02 10:31:57.231353633 +0000 UTC m=+770.561492493" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.233540 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh"] Dec 02 10:31:57 crc kubenswrapper[4679]: E1202 10:31:57.233810 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="extract" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.233829 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="extract" Dec 02 10:31:57 crc kubenswrapper[4679]: E1202 10:31:57.233846 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="pull" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.233856 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="pull" Dec 02 10:31:57 crc kubenswrapper[4679]: E1202 10:31:57.233874 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="util" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.233882 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="util" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.234027 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8d2b3f-baec-4d47-be99-1d73bc31416d" containerName="extract" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.234558 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.237544 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.238201 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.238213 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dnsbq" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.245004 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh"] Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.394916 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqcj2\" (UniqueName: \"kubernetes.io/projected/0cdf81d6-53ba-4fae-95ff-6d0d28b2384e-kube-api-access-kqcj2\") pod \"nmstate-operator-5b5b58f5c8-hnqnh\" (UID: \"0cdf81d6-53ba-4fae-95ff-6d0d28b2384e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.495934 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqcj2\" (UniqueName: \"kubernetes.io/projected/0cdf81d6-53ba-4fae-95ff-6d0d28b2384e-kube-api-access-kqcj2\") pod \"nmstate-operator-5b5b58f5c8-hnqnh\" (UID: \"0cdf81d6-53ba-4fae-95ff-6d0d28b2384e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.530168 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqcj2\" (UniqueName: \"kubernetes.io/projected/0cdf81d6-53ba-4fae-95ff-6d0d28b2384e-kube-api-access-kqcj2\") pod \"nmstate-operator-5b5b58f5c8-hnqnh\" (UID: \"0cdf81d6-53ba-4fae-95ff-6d0d28b2384e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.552130 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" Dec 02 10:31:57 crc kubenswrapper[4679]: I1202 10:31:57.733526 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh"] Dec 02 10:31:58 crc kubenswrapper[4679]: I1202 10:31:58.502292 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" event={"ID":"0cdf81d6-53ba-4fae-95ff-6d0d28b2384e","Type":"ContainerStarted","Data":"7a5bf032d9ea921b1b19881ce3c67030ba704d6a6f5ff0fd5079d1ab98d595b7"} Dec 02 10:31:59 crc kubenswrapper[4679]: I1202 10:31:59.456630 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:59 crc kubenswrapper[4679]: I1202 10:31:59.456682 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:59 crc kubenswrapper[4679]: I1202 10:31:59.499393 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:31:59 crc kubenswrapper[4679]: I1202 10:31:59.544753 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:32:01 crc kubenswrapper[4679]: I1202 10:32:01.521127 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" event={"ID":"0cdf81d6-53ba-4fae-95ff-6d0d28b2384e","Type":"ContainerStarted","Data":"9330d4b56c08c3cbe675c66990f6051855f613a997952cdb31047660cbe111d6"} Dec 02 10:32:01 crc kubenswrapper[4679]: I1202 10:32:01.545824 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-hnqnh" podStartSLOduration=1.162732917 podStartE2EDuration="4.545801332s" podCreationTimestamp="2025-12-02 10:31:57 +0000 UTC" firstStartedPulling="2025-12-02 10:31:57.740420268 +0000 UTC m=+771.070559128" lastFinishedPulling="2025-12-02 10:32:01.123488683 +0000 UTC m=+774.453627543" observedRunningTime="2025-12-02 10:32:01.54267323 +0000 UTC m=+774.872812090" watchObservedRunningTime="2025-12-02 10:32:01.545801332 +0000 UTC m=+774.875940192" Dec 02 10:32:01 crc kubenswrapper[4679]: I1202 10:32:01.726777 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-chrjz"] Dec 02 10:32:01 crc kubenswrapper[4679]: I1202 10:32:01.729696 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-chrjz" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="registry-server" containerID="cri-o://9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b" gracePeriod=2 Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.111341 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.279179 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-catalog-content\") pod \"a78093b3-ae68-4a47-a34e-57799e342169\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.279232 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-utilities\") pod \"a78093b3-ae68-4a47-a34e-57799e342169\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.279271 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtqqk\" (UniqueName: \"kubernetes.io/projected/a78093b3-ae68-4a47-a34e-57799e342169-kube-api-access-wtqqk\") pod \"a78093b3-ae68-4a47-a34e-57799e342169\" (UID: \"a78093b3-ae68-4a47-a34e-57799e342169\") " Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.280283 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-utilities" (OuterVolumeSpecName: "utilities") pod "a78093b3-ae68-4a47-a34e-57799e342169" (UID: "a78093b3-ae68-4a47-a34e-57799e342169"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.286291 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78093b3-ae68-4a47-a34e-57799e342169-kube-api-access-wtqqk" (OuterVolumeSpecName: "kube-api-access-wtqqk") pod "a78093b3-ae68-4a47-a34e-57799e342169" (UID: "a78093b3-ae68-4a47-a34e-57799e342169"). InnerVolumeSpecName "kube-api-access-wtqqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.381408 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtqqk\" (UniqueName: \"kubernetes.io/projected/a78093b3-ae68-4a47-a34e-57799e342169-kube-api-access-wtqqk\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.381442 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.530602 4679 generic.go:334] "Generic (PLEG): container finished" podID="a78093b3-ae68-4a47-a34e-57799e342169" containerID="9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b" exitCode=0 Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.530670 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chrjz" event={"ID":"a78093b3-ae68-4a47-a34e-57799e342169","Type":"ContainerDied","Data":"9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b"} Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.530721 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-chrjz" event={"ID":"a78093b3-ae68-4a47-a34e-57799e342169","Type":"ContainerDied","Data":"e7f14f0eee13208d8ba73d9c4c1ddaba84832b4ec86e2f242873fd1db1a6888d"} Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.530770 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-chrjz" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.530765 4679 scope.go:117] "RemoveContainer" containerID="9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.554804 4679 scope.go:117] "RemoveContainer" containerID="b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.576596 4679 scope.go:117] "RemoveContainer" containerID="a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.631965 4679 scope.go:117] "RemoveContainer" containerID="9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b" Dec 02 10:32:02 crc kubenswrapper[4679]: E1202 10:32:02.632536 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b\": container with ID starting with 9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b not found: ID does not exist" containerID="9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.632594 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b"} err="failed to get container status \"9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b\": rpc error: code = NotFound desc = could not find container \"9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b\": container with ID starting with 9aa5d982d896202d31dc11bb7842af3ad54d2bb114191b485bb4fa23fbbf3f8b not found: ID does not exist" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.632638 4679 scope.go:117] "RemoveContainer" containerID="b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028" Dec 02 10:32:02 crc kubenswrapper[4679]: E1202 10:32:02.633417 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028\": container with ID starting with b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028 not found: ID does not exist" containerID="b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.633467 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028"} err="failed to get container status \"b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028\": rpc error: code = NotFound desc = could not find container \"b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028\": container with ID starting with b936602f198101e335d05de58ca12263841c9dbf568d72d40d7fff2a2cc3a028 not found: ID does not exist" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.633504 4679 scope.go:117] "RemoveContainer" containerID="a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6" Dec 02 10:32:02 crc kubenswrapper[4679]: E1202 10:32:02.633849 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6\": container with ID starting with a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6 not found: ID does not exist" containerID="a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6" Dec 02 10:32:02 crc kubenswrapper[4679]: I1202 10:32:02.633881 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6"} err="failed to get container status \"a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6\": rpc error: code = NotFound desc = could not find container \"a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6\": container with ID starting with a931fd0a89732483a5fded7e5231b513571dceba8a3d60ae37fe0aa8664408b6 not found: ID does not exist" Dec 02 10:32:03 crc kubenswrapper[4679]: I1202 10:32:03.953550 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a78093b3-ae68-4a47-a34e-57799e342169" (UID: "a78093b3-ae68-4a47-a34e-57799e342169"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:32:04 crc kubenswrapper[4679]: I1202 10:32:04.003630 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78093b3-ae68-4a47-a34e-57799e342169-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:04 crc kubenswrapper[4679]: I1202 10:32:04.068936 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-chrjz"] Dec 02 10:32:04 crc kubenswrapper[4679]: I1202 10:32:04.076107 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-chrjz"] Dec 02 10:32:04 crc kubenswrapper[4679]: I1202 10:32:04.919700 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78093b3-ae68-4a47-a34e-57799e342169" path="/var/lib/kubelet/pods/a78093b3-ae68-4a47-a34e-57799e342169/volumes" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.256642 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz"] Dec 02 10:32:07 crc kubenswrapper[4679]: E1202 10:32:07.256905 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="extract-content" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.256920 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="extract-content" Dec 02 10:32:07 crc kubenswrapper[4679]: E1202 10:32:07.256933 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="registry-server" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.256941 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="registry-server" Dec 02 10:32:07 crc kubenswrapper[4679]: E1202 10:32:07.256955 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="extract-utilities" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.256963 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="extract-utilities" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.257084 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78093b3-ae68-4a47-a34e-57799e342169" containerName="registry-server" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.257775 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.259834 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xdkl6" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.285887 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.294581 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.295327 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.297289 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.332989 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-t6qm2"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.333758 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349201 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349747 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-nmstate-lock\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349789 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-ovs-socket\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349812 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntx4z\" (UniqueName: \"kubernetes.io/projected/338504ed-81c0-4c88-8d99-b1ffc9cd7bcb-kube-api-access-ntx4z\") pod \"nmstate-webhook-5f6d4c5ccb-fmbsj\" (UID: \"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349832 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vj7p\" (UniqueName: \"kubernetes.io/projected/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-kube-api-access-9vj7p\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349849 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/338504ed-81c0-4c88-8d99-b1ffc9cd7bcb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-fmbsj\" (UID: \"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349880 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hbjv\" (UniqueName: \"kubernetes.io/projected/1daefa02-7428-44b6-8da5-f1ecf0d5c12e-kube-api-access-2hbjv\") pod \"nmstate-metrics-7f946cbc9-9w2kz\" (UID: \"1daefa02-7428-44b6-8da5-f1ecf0d5c12e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.349901 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-dbus-socket\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.441501 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.442326 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.444587 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dlmtd" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.449877 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.450899 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-nmstate-lock\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.450940 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-ovs-socket\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.450982 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntx4z\" (UniqueName: \"kubernetes.io/projected/338504ed-81c0-4c88-8d99-b1ffc9cd7bcb-kube-api-access-ntx4z\") pod \"nmstate-webhook-5f6d4c5ccb-fmbsj\" (UID: \"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451013 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vj7p\" (UniqueName: \"kubernetes.io/projected/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-kube-api-access-9vj7p\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451037 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/338504ed-81c0-4c88-8d99-b1ffc9cd7bcb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-fmbsj\" (UID: \"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451083 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hbjv\" (UniqueName: \"kubernetes.io/projected/1daefa02-7428-44b6-8da5-f1ecf0d5c12e-kube-api-access-2hbjv\") pod \"nmstate-metrics-7f946cbc9-9w2kz\" (UID: \"1daefa02-7428-44b6-8da5-f1ecf0d5c12e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451113 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-dbus-socket\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451534 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-dbus-socket\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451589 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-nmstate-lock\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.451623 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-ovs-socket\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.455680 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.464706 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.473927 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/338504ed-81c0-4c88-8d99-b1ffc9cd7bcb-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-fmbsj\" (UID: \"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.478014 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntx4z\" (UniqueName: \"kubernetes.io/projected/338504ed-81c0-4c88-8d99-b1ffc9cd7bcb-kube-api-access-ntx4z\") pod \"nmstate-webhook-5f6d4c5ccb-fmbsj\" (UID: \"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.479886 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vj7p\" (UniqueName: \"kubernetes.io/projected/49ed79ab-5582-4f7e-bfd9-18263fa9ccc2-kube-api-access-9vj7p\") pod \"nmstate-handler-t6qm2\" (UID: \"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2\") " pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.492859 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hbjv\" (UniqueName: \"kubernetes.io/projected/1daefa02-7428-44b6-8da5-f1ecf0d5c12e-kube-api-access-2hbjv\") pod \"nmstate-metrics-7f946cbc9-9w2kz\" (UID: \"1daefa02-7428-44b6-8da5-f1ecf0d5c12e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.552479 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/932b0be3-5659-4b5b-a667-cd03f56c8f98-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.552604 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/932b0be3-5659-4b5b-a667-cd03f56c8f98-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.552653 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7zgs\" (UniqueName: \"kubernetes.io/projected/932b0be3-5659-4b5b-a667-cd03f56c8f98-kube-api-access-b7zgs\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.576320 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.637852 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.654276 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/932b0be3-5659-4b5b-a667-cd03f56c8f98-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.654372 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7zgs\" (UniqueName: \"kubernetes.io/projected/932b0be3-5659-4b5b-a667-cd03f56c8f98-kube-api-access-b7zgs\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.654409 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.654462 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/932b0be3-5659-4b5b-a667-cd03f56c8f98-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: E1202 10:32:07.654989 4679 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 02 10:32:07 crc kubenswrapper[4679]: E1202 10:32:07.655056 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/932b0be3-5659-4b5b-a667-cd03f56c8f98-plugin-serving-cert podName:932b0be3-5659-4b5b-a667-cd03f56c8f98 nodeName:}" failed. No retries permitted until 2025-12-02 10:32:08.155034598 +0000 UTC m=+781.485173508 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/932b0be3-5659-4b5b-a667-cd03f56c8f98-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-r2xnw" (UID: "932b0be3-5659-4b5b-a667-cd03f56c8f98") : secret "plugin-serving-cert" not found Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.655444 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/932b0be3-5659-4b5b-a667-cd03f56c8f98-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.666731 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-77c4f49446-ss295"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.667503 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.677724 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7zgs\" (UniqueName: \"kubernetes.io/projected/932b0be3-5659-4b5b-a667-cd03f56c8f98-kube-api-access-b7zgs\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.683603 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77c4f49446-ss295"] Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.857248 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-service-ca\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.858391 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znj5t\" (UniqueName: \"kubernetes.io/projected/e01142d9-93ce-41a3-9160-c2c46e5f64e8-kube-api-access-znj5t\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.858445 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-oauth-config\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.858510 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-config\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.858589 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-oauth-serving-cert\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.858643 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-trusted-ca-bundle\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.858687 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-serving-cert\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.899008 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj"] Dec 02 10:32:07 crc kubenswrapper[4679]: W1202 10:32:07.902804 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod338504ed_81c0_4c88_8d99_b1ffc9cd7bcb.slice/crio-f1c80aae0f769db62dd0c62fff9dc66d7e5acf6601d427aaebf5b4f0093eb0c2 WatchSource:0}: Error finding container f1c80aae0f769db62dd0c62fff9dc66d7e5acf6601d427aaebf5b4f0093eb0c2: Status 404 returned error can't find the container with id f1c80aae0f769db62dd0c62fff9dc66d7e5acf6601d427aaebf5b4f0093eb0c2 Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959703 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-service-ca\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959793 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znj5t\" (UniqueName: \"kubernetes.io/projected/e01142d9-93ce-41a3-9160-c2c46e5f64e8-kube-api-access-znj5t\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959824 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-oauth-config\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959858 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-config\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959874 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-oauth-serving-cert\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959890 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-trusted-ca-bundle\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.959916 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-serving-cert\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.961143 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-service-ca\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.961171 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-config\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.962085 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-trusted-ca-bundle\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.962098 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e01142d9-93ce-41a3-9160-c2c46e5f64e8-oauth-serving-cert\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.964005 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-oauth-config\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.964651 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e01142d9-93ce-41a3-9160-c2c46e5f64e8-console-serving-cert\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.982229 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znj5t\" (UniqueName: \"kubernetes.io/projected/e01142d9-93ce-41a3-9160-c2c46e5f64e8-kube-api-access-znj5t\") pod \"console-77c4f49446-ss295\" (UID: \"e01142d9-93ce-41a3-9160-c2c46e5f64e8\") " pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:07 crc kubenswrapper[4679]: I1202 10:32:07.996465 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.052207 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz"] Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.162651 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/932b0be3-5659-4b5b-a667-cd03f56c8f98-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.169051 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/932b0be3-5659-4b5b-a667-cd03f56c8f98-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-r2xnw\" (UID: \"932b0be3-5659-4b5b-a667-cd03f56c8f98\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.214047 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77c4f49446-ss295"] Dec 02 10:32:08 crc kubenswrapper[4679]: W1202 10:32:08.222527 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode01142d9_93ce_41a3_9160_c2c46e5f64e8.slice/crio-fd7b1361417f8d90a3b16bfa3034d5f2d992ab5b36f3c6d075e21192b74e9d1b WatchSource:0}: Error finding container fd7b1361417f8d90a3b16bfa3034d5f2d992ab5b36f3c6d075e21192b74e9d1b: Status 404 returned error can't find the container with id fd7b1361417f8d90a3b16bfa3034d5f2d992ab5b36f3c6d075e21192b74e9d1b Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.419922 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.583042 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-t6qm2" event={"ID":"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2","Type":"ContainerStarted","Data":"c90432b590f086b6efbc023da4b19e073ab4965184b925fc2c7448d9d86d732a"} Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.586455 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" event={"ID":"1daefa02-7428-44b6-8da5-f1ecf0d5c12e","Type":"ContainerStarted","Data":"e5cf62a56d77e5baa5677dc4077f0865fd4d14454c710fd069645425683f3285"} Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.612500 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77c4f49446-ss295" event={"ID":"e01142d9-93ce-41a3-9160-c2c46e5f64e8","Type":"ContainerStarted","Data":"0eaf12aeb3ad54c497edd13eeb94af4c17e6d941beaf3b030d8a4712fe2af502"} Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.612656 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77c4f49446-ss295" event={"ID":"e01142d9-93ce-41a3-9160-c2c46e5f64e8","Type":"ContainerStarted","Data":"fd7b1361417f8d90a3b16bfa3034d5f2d992ab5b36f3c6d075e21192b74e9d1b"} Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.617269 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" event={"ID":"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb","Type":"ContainerStarted","Data":"f1c80aae0f769db62dd0c62fff9dc66d7e5acf6601d427aaebf5b4f0093eb0c2"} Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.634329 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-77c4f49446-ss295" podStartSLOduration=1.634282517 podStartE2EDuration="1.634282517s" podCreationTimestamp="2025-12-02 10:32:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:32:08.633050571 +0000 UTC m=+781.963189441" watchObservedRunningTime="2025-12-02 10:32:08.634282517 +0000 UTC m=+781.964421397" Dec 02 10:32:08 crc kubenswrapper[4679]: W1202 10:32:08.670603 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod932b0be3_5659_4b5b_a667_cd03f56c8f98.slice/crio-5d1e28bf1e247ed6986fdc9506bd8ece597b32ff0475d13d3354aca1b65cc4a4 WatchSource:0}: Error finding container 5d1e28bf1e247ed6986fdc9506bd8ece597b32ff0475d13d3354aca1b65cc4a4: Status 404 returned error can't find the container with id 5d1e28bf1e247ed6986fdc9506bd8ece597b32ff0475d13d3354aca1b65cc4a4 Dec 02 10:32:08 crc kubenswrapper[4679]: I1202 10:32:08.673550 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw"] Dec 02 10:32:09 crc kubenswrapper[4679]: I1202 10:32:09.626026 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" event={"ID":"932b0be3-5659-4b5b-a667-cd03f56c8f98","Type":"ContainerStarted","Data":"5d1e28bf1e247ed6986fdc9506bd8ece597b32ff0475d13d3354aca1b65cc4a4"} Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.640781 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" event={"ID":"1daefa02-7428-44b6-8da5-f1ecf0d5c12e","Type":"ContainerStarted","Data":"2dd9208bb06a43915c19e33508a5525134bb5502af80452ac66ee776109a0f62"} Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.642721 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" event={"ID":"338504ed-81c0-4c88-8d99-b1ffc9cd7bcb","Type":"ContainerStarted","Data":"eb100e3578908f6a33153284494f1daf1632eed2ed5d17b5561495976b3032d0"} Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.644448 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.647741 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-t6qm2" event={"ID":"49ed79ab-5582-4f7e-bfd9-18263fa9ccc2","Type":"ContainerStarted","Data":"8ad53e967b148ff8829581f0933469c58f10ae301b52dfbc10a8cc829228be85"} Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.647934 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.687829 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" podStartSLOduration=1.9949860799999999 podStartE2EDuration="4.687807091s" podCreationTimestamp="2025-12-02 10:32:07 +0000 UTC" firstStartedPulling="2025-12-02 10:32:07.905437324 +0000 UTC m=+781.235576184" lastFinishedPulling="2025-12-02 10:32:10.598258335 +0000 UTC m=+783.928397195" observedRunningTime="2025-12-02 10:32:11.66225024 +0000 UTC m=+784.992389120" watchObservedRunningTime="2025-12-02 10:32:11.687807091 +0000 UTC m=+785.017945971" Dec 02 10:32:11 crc kubenswrapper[4679]: I1202 10:32:11.690314 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-t6qm2" podStartSLOduration=1.683482023 podStartE2EDuration="4.690284023s" podCreationTimestamp="2025-12-02 10:32:07 +0000 UTC" firstStartedPulling="2025-12-02 10:32:07.717528029 +0000 UTC m=+781.047666889" lastFinishedPulling="2025-12-02 10:32:10.724330029 +0000 UTC m=+784.054468889" observedRunningTime="2025-12-02 10:32:11.684957598 +0000 UTC m=+785.015096458" watchObservedRunningTime="2025-12-02 10:32:11.690284023 +0000 UTC m=+785.020422883" Dec 02 10:32:12 crc kubenswrapper[4679]: I1202 10:32:12.655573 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" event={"ID":"932b0be3-5659-4b5b-a667-cd03f56c8f98","Type":"ContainerStarted","Data":"392768d16788544e74b7e3c2a5298cf109210c03067998dccaf087ae45b19779"} Dec 02 10:32:12 crc kubenswrapper[4679]: I1202 10:32:12.676104 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-r2xnw" podStartSLOduration=2.717020296 podStartE2EDuration="5.676080262s" podCreationTimestamp="2025-12-02 10:32:07 +0000 UTC" firstStartedPulling="2025-12-02 10:32:08.673107742 +0000 UTC m=+782.003246602" lastFinishedPulling="2025-12-02 10:32:11.632167708 +0000 UTC m=+784.962306568" observedRunningTime="2025-12-02 10:32:12.671345565 +0000 UTC m=+786.001484445" watchObservedRunningTime="2025-12-02 10:32:12.676080262 +0000 UTC m=+786.006219122" Dec 02 10:32:13 crc kubenswrapper[4679]: I1202 10:32:13.664380 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" event={"ID":"1daefa02-7428-44b6-8da5-f1ecf0d5c12e","Type":"ContainerStarted","Data":"bdd60627785f2673f3473bf816bd99beea246a2e184c7d082c053fb74edf86e9"} Dec 02 10:32:13 crc kubenswrapper[4679]: I1202 10:32:13.686064 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9w2kz" podStartSLOduration=1.593151555 podStartE2EDuration="6.686044452s" podCreationTimestamp="2025-12-02 10:32:07 +0000 UTC" firstStartedPulling="2025-12-02 10:32:08.065943426 +0000 UTC m=+781.396082286" lastFinishedPulling="2025-12-02 10:32:13.158836313 +0000 UTC m=+786.488975183" observedRunningTime="2025-12-02 10:32:13.680991285 +0000 UTC m=+787.011130165" watchObservedRunningTime="2025-12-02 10:32:13.686044452 +0000 UTC m=+787.016183312" Dec 02 10:32:16 crc kubenswrapper[4679]: I1202 10:32:16.932389 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:32:16 crc kubenswrapper[4679]: I1202 10:32:16.932745 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:32:17 crc kubenswrapper[4679]: I1202 10:32:17.679859 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-t6qm2" Dec 02 10:32:17 crc kubenswrapper[4679]: I1202 10:32:17.997103 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:17 crc kubenswrapper[4679]: I1202 10:32:17.997165 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:18 crc kubenswrapper[4679]: I1202 10:32:18.006124 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:18 crc kubenswrapper[4679]: I1202 10:32:18.696972 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-77c4f49446-ss295" Dec 02 10:32:18 crc kubenswrapper[4679]: I1202 10:32:18.772437 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5bk2f"] Dec 02 10:32:27 crc kubenswrapper[4679]: I1202 10:32:27.646519 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-fmbsj" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.559432 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm"] Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.561033 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.562836 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.573480 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm"] Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.684969 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.685187 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.685500 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f947q\" (UniqueName: \"kubernetes.io/projected/646ec543-e35d-4099-bf1a-d0284276fc95-kube-api-access-f947q\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.786664 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.787093 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.787409 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f947q\" (UniqueName: \"kubernetes.io/projected/646ec543-e35d-4099-bf1a-d0284276fc95-kube-api-access-f947q\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.787816 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.787819 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.818136 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5bk2f" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerName="console" containerID="cri-o://1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892" gracePeriod=15 Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.823673 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f947q\" (UniqueName: \"kubernetes.io/projected/646ec543-e35d-4099-bf1a-d0284276fc95-kube-api-access-f947q\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:43 crc kubenswrapper[4679]: I1202 10:32:43.885101 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.242619 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5bk2f_9d97c73c-99b7-452c-9e31-f7c99d4c93f3/console/0.log" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.242689 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294118 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-oauth-serving-cert\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294247 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-serving-cert\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294296 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-config\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294398 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-oauth-config\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294422 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-trusted-ca-bundle\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294442 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-service-ca\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.294480 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfpt6\" (UniqueName: \"kubernetes.io/projected/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-kube-api-access-kfpt6\") pod \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\" (UID: \"9d97c73c-99b7-452c-9e31-f7c99d4c93f3\") " Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.295146 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-service-ca" (OuterVolumeSpecName: "service-ca") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.295379 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.295404 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.295461 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-config" (OuterVolumeSpecName: "console-config") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.299810 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-kube-api-access-kfpt6" (OuterVolumeSpecName: "kube-api-access-kfpt6") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "kube-api-access-kfpt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.300045 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.300663 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9d97c73c-99b7-452c-9e31-f7c99d4c93f3" (UID: "9d97c73c-99b7-452c-9e31-f7c99d4c93f3"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.334212 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm"] Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395775 4679 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395797 4679 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395806 4679 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395814 4679 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395822 4679 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395831 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfpt6\" (UniqueName: \"kubernetes.io/projected/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-kube-api-access-kfpt6\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.395839 4679 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d97c73c-99b7-452c-9e31-f7c99d4c93f3-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.849375 4679 generic.go:334] "Generic (PLEG): container finished" podID="646ec543-e35d-4099-bf1a-d0284276fc95" containerID="95fb512e91a18789d69cbc520aeb8ced4d0ff4f6831b9a1b0a0dc44e29a50866" exitCode=0 Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.849462 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" event={"ID":"646ec543-e35d-4099-bf1a-d0284276fc95","Type":"ContainerDied","Data":"95fb512e91a18789d69cbc520aeb8ced4d0ff4f6831b9a1b0a0dc44e29a50866"} Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.849523 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" event={"ID":"646ec543-e35d-4099-bf1a-d0284276fc95","Type":"ContainerStarted","Data":"aa732df1410e3fa54a5cfbabd0fdc446f554d21f36d34199a266d5a23f798dcb"} Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.851014 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5bk2f_9d97c73c-99b7-452c-9e31-f7c99d4c93f3/console/0.log" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.851085 4679 generic.go:334] "Generic (PLEG): container finished" podID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerID="1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892" exitCode=2 Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.851112 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bk2f" event={"ID":"9d97c73c-99b7-452c-9e31-f7c99d4c93f3","Type":"ContainerDied","Data":"1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892"} Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.851147 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bk2f" event={"ID":"9d97c73c-99b7-452c-9e31-f7c99d4c93f3","Type":"ContainerDied","Data":"b99003e690139e98d2c252d58c2dde8dfbc7bf76b57f3438f9da61bce9a1a90a"} Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.851163 4679 scope.go:117] "RemoveContainer" containerID="1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.851207 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bk2f" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.868942 4679 scope.go:117] "RemoveContainer" containerID="1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892" Dec 02 10:32:44 crc kubenswrapper[4679]: E1202 10:32:44.869266 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892\": container with ID starting with 1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892 not found: ID does not exist" containerID="1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.869294 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892"} err="failed to get container status \"1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892\": rpc error: code = NotFound desc = could not find container \"1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892\": container with ID starting with 1c9ade6f459943a0ebd8d633ba78bd41ac5e89662d9a9a0576e215d6920f3892 not found: ID does not exist" Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.897228 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5bk2f"] Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.902237 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5bk2f"] Dec 02 10:32:44 crc kubenswrapper[4679]: I1202 10:32:44.917253 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" path="/var/lib/kubelet/pods/9d97c73c-99b7-452c-9e31-f7c99d4c93f3/volumes" Dec 02 10:32:46 crc kubenswrapper[4679]: I1202 10:32:46.867093 4679 generic.go:334] "Generic (PLEG): container finished" podID="646ec543-e35d-4099-bf1a-d0284276fc95" containerID="d88b325c1f3826c2e37adf26ad7e6a6c46ecdd2bffce17dde4bb09d65062c65e" exitCode=0 Dec 02 10:32:46 crc kubenswrapper[4679]: I1202 10:32:46.867260 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" event={"ID":"646ec543-e35d-4099-bf1a-d0284276fc95","Type":"ContainerDied","Data":"d88b325c1f3826c2e37adf26ad7e6a6c46ecdd2bffce17dde4bb09d65062c65e"} Dec 02 10:32:46 crc kubenswrapper[4679]: I1202 10:32:46.932496 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:32:46 crc kubenswrapper[4679]: I1202 10:32:46.932580 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:32:47 crc kubenswrapper[4679]: I1202 10:32:47.875939 4679 generic.go:334] "Generic (PLEG): container finished" podID="646ec543-e35d-4099-bf1a-d0284276fc95" containerID="3ea186b8668184a77471d63cad24621345ccb7f07920225e9b506f376532436d" exitCode=0 Dec 02 10:32:47 crc kubenswrapper[4679]: I1202 10:32:47.875976 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" event={"ID":"646ec543-e35d-4099-bf1a-d0284276fc95","Type":"ContainerDied","Data":"3ea186b8668184a77471d63cad24621345ccb7f07920225e9b506f376532436d"} Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.190605 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.256353 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f947q\" (UniqueName: \"kubernetes.io/projected/646ec543-e35d-4099-bf1a-d0284276fc95-kube-api-access-f947q\") pod \"646ec543-e35d-4099-bf1a-d0284276fc95\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.256433 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-bundle\") pod \"646ec543-e35d-4099-bf1a-d0284276fc95\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.256498 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-util\") pod \"646ec543-e35d-4099-bf1a-d0284276fc95\" (UID: \"646ec543-e35d-4099-bf1a-d0284276fc95\") " Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.257895 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-bundle" (OuterVolumeSpecName: "bundle") pod "646ec543-e35d-4099-bf1a-d0284276fc95" (UID: "646ec543-e35d-4099-bf1a-d0284276fc95"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.259027 4679 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.262570 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/646ec543-e35d-4099-bf1a-d0284276fc95-kube-api-access-f947q" (OuterVolumeSpecName: "kube-api-access-f947q") pod "646ec543-e35d-4099-bf1a-d0284276fc95" (UID: "646ec543-e35d-4099-bf1a-d0284276fc95"). InnerVolumeSpecName "kube-api-access-f947q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.270460 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-util" (OuterVolumeSpecName: "util") pod "646ec543-e35d-4099-bf1a-d0284276fc95" (UID: "646ec543-e35d-4099-bf1a-d0284276fc95"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.360508 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f947q\" (UniqueName: \"kubernetes.io/projected/646ec543-e35d-4099-bf1a-d0284276fc95-kube-api-access-f947q\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.360545 4679 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/646ec543-e35d-4099-bf1a-d0284276fc95-util\") on node \"crc\" DevicePath \"\"" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.892800 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" event={"ID":"646ec543-e35d-4099-bf1a-d0284276fc95","Type":"ContainerDied","Data":"aa732df1410e3fa54a5cfbabd0fdc446f554d21f36d34199a266d5a23f798dcb"} Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.892886 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa732df1410e3fa54a5cfbabd0fdc446f554d21f36d34199a266d5a23f798dcb" Dec 02 10:32:49 crc kubenswrapper[4679]: I1202 10:32:49.892844 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.403495 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj"] Dec 02 10:32:58 crc kubenswrapper[4679]: E1202 10:32:58.404266 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="util" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.404283 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="util" Dec 02 10:32:58 crc kubenswrapper[4679]: E1202 10:32:58.404322 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerName="console" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.404330 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerName="console" Dec 02 10:32:58 crc kubenswrapper[4679]: E1202 10:32:58.404342 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="pull" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.404350 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="pull" Dec 02 10:32:58 crc kubenswrapper[4679]: E1202 10:32:58.404372 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="extract" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.404379 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="extract" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.404513 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d97c73c-99b7-452c-9e31-f7c99d4c93f3" containerName="console" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.404526 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="646ec543-e35d-4099-bf1a-d0284276fc95" containerName="extract" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.405006 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.408950 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.408970 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.409080 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-56k5q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.409171 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.409209 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.424099 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj"] Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.492808 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bltsj\" (UniqueName: \"kubernetes.io/projected/06025ad8-fd55-4fe1-95bd-08a0c83168e0-kube-api-access-bltsj\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.493378 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06025ad8-fd55-4fe1-95bd-08a0c83168e0-apiservice-cert\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.493526 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06025ad8-fd55-4fe1-95bd-08a0c83168e0-webhook-cert\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.595431 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bltsj\" (UniqueName: \"kubernetes.io/projected/06025ad8-fd55-4fe1-95bd-08a0c83168e0-kube-api-access-bltsj\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.595715 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06025ad8-fd55-4fe1-95bd-08a0c83168e0-apiservice-cert\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.595801 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06025ad8-fd55-4fe1-95bd-08a0c83168e0-webhook-cert\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.605150 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06025ad8-fd55-4fe1-95bd-08a0c83168e0-apiservice-cert\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.605367 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06025ad8-fd55-4fe1-95bd-08a0c83168e0-webhook-cert\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.620462 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bltsj\" (UniqueName: \"kubernetes.io/projected/06025ad8-fd55-4fe1-95bd-08a0c83168e0-kube-api-access-bltsj\") pod \"metallb-operator-controller-manager-8555c5bdfd-j4hqj\" (UID: \"06025ad8-fd55-4fe1-95bd-08a0c83168e0\") " pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.672148 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q"] Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.672957 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.676663 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.676914 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.677968 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rc2gf" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.730824 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.738451 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q"] Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.798263 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4086bfe3-e82b-46c4-87f6-837a42348472-apiservice-cert\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.798475 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4086bfe3-e82b-46c4-87f6-837a42348472-webhook-cert\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.798560 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wmsm\" (UniqueName: \"kubernetes.io/projected/4086bfe3-e82b-46c4-87f6-837a42348472-kube-api-access-6wmsm\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.899905 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wmsm\" (UniqueName: \"kubernetes.io/projected/4086bfe3-e82b-46c4-87f6-837a42348472-kube-api-access-6wmsm\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.900251 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4086bfe3-e82b-46c4-87f6-837a42348472-apiservice-cert\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.900288 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4086bfe3-e82b-46c4-87f6-837a42348472-webhook-cert\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.904912 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4086bfe3-e82b-46c4-87f6-837a42348472-apiservice-cert\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.910946 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4086bfe3-e82b-46c4-87f6-837a42348472-webhook-cert\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:58 crc kubenswrapper[4679]: I1202 10:32:58.923095 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wmsm\" (UniqueName: \"kubernetes.io/projected/4086bfe3-e82b-46c4-87f6-837a42348472-kube-api-access-6wmsm\") pod \"metallb-operator-webhook-server-898774cd7-nvg9q\" (UID: \"4086bfe3-e82b-46c4-87f6-837a42348472\") " pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:59 crc kubenswrapper[4679]: I1202 10:32:59.039475 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:32:59 crc kubenswrapper[4679]: I1202 10:32:59.127071 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj"] Dec 02 10:32:59 crc kubenswrapper[4679]: W1202 10:32:59.148507 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06025ad8_fd55_4fe1_95bd_08a0c83168e0.slice/crio-e9799acab59ae4abaa607d1e09743c2be84dd83529ce83d09bfb0ed99eb7062c WatchSource:0}: Error finding container e9799acab59ae4abaa607d1e09743c2be84dd83529ce83d09bfb0ed99eb7062c: Status 404 returned error can't find the container with id e9799acab59ae4abaa607d1e09743c2be84dd83529ce83d09bfb0ed99eb7062c Dec 02 10:32:59 crc kubenswrapper[4679]: I1202 10:32:59.477689 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q"] Dec 02 10:32:59 crc kubenswrapper[4679]: W1202 10:32:59.482023 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4086bfe3_e82b_46c4_87f6_837a42348472.slice/crio-26f7cb67579e04190730afc429fdc7edadb2086faae10797e3af4c61dcb5e88e WatchSource:0}: Error finding container 26f7cb67579e04190730afc429fdc7edadb2086faae10797e3af4c61dcb5e88e: Status 404 returned error can't find the container with id 26f7cb67579e04190730afc429fdc7edadb2086faae10797e3af4c61dcb5e88e Dec 02 10:32:59 crc kubenswrapper[4679]: I1202 10:32:59.958082 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" event={"ID":"4086bfe3-e82b-46c4-87f6-837a42348472","Type":"ContainerStarted","Data":"26f7cb67579e04190730afc429fdc7edadb2086faae10797e3af4c61dcb5e88e"} Dec 02 10:32:59 crc kubenswrapper[4679]: I1202 10:32:59.958896 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" event={"ID":"06025ad8-fd55-4fe1-95bd-08a0c83168e0","Type":"ContainerStarted","Data":"e9799acab59ae4abaa607d1e09743c2be84dd83529ce83d09bfb0ed99eb7062c"} Dec 02 10:33:02 crc kubenswrapper[4679]: I1202 10:33:02.984145 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" event={"ID":"06025ad8-fd55-4fe1-95bd-08a0c83168e0","Type":"ContainerStarted","Data":"714057c2722134d9556b379bb5be894f1f21d12856551e19257b7e94a2b1a504"} Dec 02 10:33:02 crc kubenswrapper[4679]: I1202 10:33:02.984259 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:33:03 crc kubenswrapper[4679]: I1202 10:33:03.014286 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" podStartSLOduration=2.191457739 podStartE2EDuration="5.014260926s" podCreationTimestamp="2025-12-02 10:32:58 +0000 UTC" firstStartedPulling="2025-12-02 10:32:59.150778519 +0000 UTC m=+832.480917379" lastFinishedPulling="2025-12-02 10:33:01.973581706 +0000 UTC m=+835.303720566" observedRunningTime="2025-12-02 10:33:03.011041212 +0000 UTC m=+836.341180062" watchObservedRunningTime="2025-12-02 10:33:03.014260926 +0000 UTC m=+836.344399786" Dec 02 10:33:05 crc kubenswrapper[4679]: I1202 10:33:04.999958 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" event={"ID":"4086bfe3-e82b-46c4-87f6-837a42348472","Type":"ContainerStarted","Data":"52ce879090218ffa5ed497e2dc86c671d0c287af3953c77a50a17716b7f835c7"} Dec 02 10:33:05 crc kubenswrapper[4679]: I1202 10:33:05.000393 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:33:05 crc kubenswrapper[4679]: I1202 10:33:05.025400 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" podStartSLOduration=2.476514651 podStartE2EDuration="7.025381661s" podCreationTimestamp="2025-12-02 10:32:58 +0000 UTC" firstStartedPulling="2025-12-02 10:32:59.485627353 +0000 UTC m=+832.815766213" lastFinishedPulling="2025-12-02 10:33:04.034494363 +0000 UTC m=+837.364633223" observedRunningTime="2025-12-02 10:33:05.0236271 +0000 UTC m=+838.353765960" watchObservedRunningTime="2025-12-02 10:33:05.025381661 +0000 UTC m=+838.355520531" Dec 02 10:33:16 crc kubenswrapper[4679]: I1202 10:33:16.931539 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:33:16 crc kubenswrapper[4679]: I1202 10:33:16.932139 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:33:16 crc kubenswrapper[4679]: I1202 10:33:16.932181 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:33:16 crc kubenswrapper[4679]: I1202 10:33:16.932764 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09c15e7820d4c5f93b890b6f7afc7d06c4c33bed2214883f94f5eae437e987f2"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:33:16 crc kubenswrapper[4679]: I1202 10:33:16.932808 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://09c15e7820d4c5f93b890b6f7afc7d06c4c33bed2214883f94f5eae437e987f2" gracePeriod=600 Dec 02 10:33:18 crc kubenswrapper[4679]: I1202 10:33:18.080009 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="09c15e7820d4c5f93b890b6f7afc7d06c4c33bed2214883f94f5eae437e987f2" exitCode=0 Dec 02 10:33:18 crc kubenswrapper[4679]: I1202 10:33:18.080084 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"09c15e7820d4c5f93b890b6f7afc7d06c4c33bed2214883f94f5eae437e987f2"} Dec 02 10:33:18 crc kubenswrapper[4679]: I1202 10:33:18.080430 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"d0ace29b98cd54337ce3c65db2564b1430008dec833a4152e92f6248a82bf4d3"} Dec 02 10:33:18 crc kubenswrapper[4679]: I1202 10:33:18.080472 4679 scope.go:117] "RemoveContainer" containerID="366ff952d0a78699e9a3954a944a0a1122afa8431e83a016c9f4d6d35d914975" Dec 02 10:33:19 crc kubenswrapper[4679]: I1202 10:33:19.046866 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-898774cd7-nvg9q" Dec 02 10:33:38 crc kubenswrapper[4679]: I1202 10:33:38.734885 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8555c5bdfd-j4hqj" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.541917 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5"] Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.543477 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.547081 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-qx64n"] Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.551505 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.554673 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.554864 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.554946 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-np7tz" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.554953 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.560484 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5"] Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567628 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-conf\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567687 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/280f2b1e-4e5a-481d-ad2f-4536d832697a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567718 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgv4n\" (UniqueName: \"kubernetes.io/projected/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-kube-api-access-zgv4n\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567746 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-sockets\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567769 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-startup\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567920 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-metrics-certs\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.567981 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-metrics\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.568016 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhkk\" (UniqueName: \"kubernetes.io/projected/280f2b1e-4e5a-481d-ad2f-4536d832697a-kube-api-access-dkhkk\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.568041 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-reloader\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669041 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-sockets\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669204 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-startup\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669265 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-metrics-certs\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669323 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-metrics\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669350 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-reloader\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669373 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhkk\" (UniqueName: \"kubernetes.io/projected/280f2b1e-4e5a-481d-ad2f-4536d832697a-kube-api-access-dkhkk\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669419 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-conf\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669460 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/280f2b1e-4e5a-481d-ad2f-4536d832697a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669487 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgv4n\" (UniqueName: \"kubernetes.io/projected/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-kube-api-access-zgv4n\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669514 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-sockets\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: E1202 10:33:39.669749 4679 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.669787 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-metrics\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: E1202 10:33:39.669871 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/280f2b1e-4e5a-481d-ad2f-4536d832697a-cert podName:280f2b1e-4e5a-481d-ad2f-4536d832697a nodeName:}" failed. No retries permitted until 2025-12-02 10:33:40.169835627 +0000 UTC m=+873.499974487 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/280f2b1e-4e5a-481d-ad2f-4536d832697a-cert") pod "frr-k8s-webhook-server-7fcb986d4-zcqd5" (UID: "280f2b1e-4e5a-481d-ad2f-4536d832697a") : secret "frr-k8s-webhook-server-cert" not found Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.670001 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-conf\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.670265 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-reloader\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.670380 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-frr-startup\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.677231 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-metrics-certs\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.706082 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgv4n\" (UniqueName: \"kubernetes.io/projected/c5c9e3af-4021-4100-8a4f-3523f5da7bf9-kube-api-access-zgv4n\") pod \"frr-k8s-qx64n\" (UID: \"c5c9e3af-4021-4100-8a4f-3523f5da7bf9\") " pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.706154 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhkk\" (UniqueName: \"kubernetes.io/projected/280f2b1e-4e5a-481d-ad2f-4536d832697a-kube-api-access-dkhkk\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.754833 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qgdwn"] Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.763416 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.768969 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.769379 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-6m7qk" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.769629 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.770396 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.771200 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.771261 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-975r2\" (UniqueName: \"kubernetes.io/projected/c42448eb-c465-4172-9830-37000ff7e881-kube-api-access-975r2\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.771384 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-metrics-certs\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.771412 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c42448eb-c465-4172-9830-37000ff7e881-metallb-excludel2\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.795963 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-ws8bf"] Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.796954 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.804681 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.822135 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-ws8bf"] Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.872909 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2302ab5-3f22-4f2e-804f-a9e1109446a6-metrics-certs\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.872987 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.873209 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-metrics-certs\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.873284 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c42448eb-c465-4172-9830-37000ff7e881-metallb-excludel2\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.873396 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.873448 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6br9q\" (UniqueName: \"kubernetes.io/projected/d2302ab5-3f22-4f2e-804f-a9e1109446a6-kube-api-access-6br9q\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.873493 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2302ab5-3f22-4f2e-804f-a9e1109446a6-cert\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.873532 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-975r2\" (UniqueName: \"kubernetes.io/projected/c42448eb-c465-4172-9830-37000ff7e881-kube-api-access-975r2\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: E1202 10:33:39.873783 4679 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 02 10:33:39 crc kubenswrapper[4679]: E1202 10:33:39.873833 4679 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 02 10:33:39 crc kubenswrapper[4679]: E1202 10:33:39.873846 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist podName:c42448eb-c465-4172-9830-37000ff7e881 nodeName:}" failed. No retries permitted until 2025-12-02 10:33:40.373825079 +0000 UTC m=+873.703963939 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist") pod "speaker-qgdwn" (UID: "c42448eb-c465-4172-9830-37000ff7e881") : secret "metallb-memberlist" not found Dec 02 10:33:39 crc kubenswrapper[4679]: E1202 10:33:39.873947 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-metrics-certs podName:c42448eb-c465-4172-9830-37000ff7e881 nodeName:}" failed. No retries permitted until 2025-12-02 10:33:40.373922352 +0000 UTC m=+873.704061292 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-metrics-certs") pod "speaker-qgdwn" (UID: "c42448eb-c465-4172-9830-37000ff7e881") : secret "speaker-certs-secret" not found Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.874699 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c42448eb-c465-4172-9830-37000ff7e881-metallb-excludel2\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.900665 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-975r2\" (UniqueName: \"kubernetes.io/projected/c42448eb-c465-4172-9830-37000ff7e881-kube-api-access-975r2\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.974393 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6br9q\" (UniqueName: \"kubernetes.io/projected/d2302ab5-3f22-4f2e-804f-a9e1109446a6-kube-api-access-6br9q\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.974437 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2302ab5-3f22-4f2e-804f-a9e1109446a6-cert\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.974825 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2302ab5-3f22-4f2e-804f-a9e1109446a6-metrics-certs\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.978380 4679 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.984944 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2302ab5-3f22-4f2e-804f-a9e1109446a6-metrics-certs\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.989368 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2302ab5-3f22-4f2e-804f-a9e1109446a6-cert\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:39 crc kubenswrapper[4679]: I1202 10:33:39.993095 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6br9q\" (UniqueName: \"kubernetes.io/projected/d2302ab5-3f22-4f2e-804f-a9e1109446a6-kube-api-access-6br9q\") pod \"controller-f8648f98b-ws8bf\" (UID: \"d2302ab5-3f22-4f2e-804f-a9e1109446a6\") " pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.126840 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.177501 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/280f2b1e-4e5a-481d-ad2f-4536d832697a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.181914 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/280f2b1e-4e5a-481d-ad2f-4536d832697a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-zcqd5\" (UID: \"280f2b1e-4e5a-481d-ad2f-4536d832697a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.224355 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"4cbb830d13fdb4163227b3164cb9a40d8b777a74a199a86a84f8faf8c4552273"} Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.381038 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-metrics-certs\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.381432 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:40 crc kubenswrapper[4679]: E1202 10:33:40.381559 4679 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 02 10:33:40 crc kubenswrapper[4679]: E1202 10:33:40.381618 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist podName:c42448eb-c465-4172-9830-37000ff7e881 nodeName:}" failed. No retries permitted until 2025-12-02 10:33:41.381599625 +0000 UTC m=+874.711738485 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist") pod "speaker-qgdwn" (UID: "c42448eb-c465-4172-9830-37000ff7e881") : secret "metallb-memberlist" not found Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.385840 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-metrics-certs\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.462288 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.566945 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-ws8bf"] Dec 02 10:33:40 crc kubenswrapper[4679]: W1202 10:33:40.575051 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2302ab5_3f22_4f2e_804f_a9e1109446a6.slice/crio-054a44c2eb52b525fe52381ca1659cf6994abe9c4b8ca2c1794ac72233ce0dfe WatchSource:0}: Error finding container 054a44c2eb52b525fe52381ca1659cf6994abe9c4b8ca2c1794ac72233ce0dfe: Status 404 returned error can't find the container with id 054a44c2eb52b525fe52381ca1659cf6994abe9c4b8ca2c1794ac72233ce0dfe Dec 02 10:33:40 crc kubenswrapper[4679]: I1202 10:33:40.663194 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5"] Dec 02 10:33:40 crc kubenswrapper[4679]: W1202 10:33:40.669845 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod280f2b1e_4e5a_481d_ad2f_4536d832697a.slice/crio-f338a7e243ee76e7d4e8d26f47cf320bf4e04a8a4994b02b2a4aa2c2b544c8cf WatchSource:0}: Error finding container f338a7e243ee76e7d4e8d26f47cf320bf4e04a8a4994b02b2a4aa2c2b544c8cf: Status 404 returned error can't find the container with id f338a7e243ee76e7d4e8d26f47cf320bf4e04a8a4994b02b2a4aa2c2b544c8cf Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.234267 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" event={"ID":"280f2b1e-4e5a-481d-ad2f-4536d832697a","Type":"ContainerStarted","Data":"f338a7e243ee76e7d4e8d26f47cf320bf4e04a8a4994b02b2a4aa2c2b544c8cf"} Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.236103 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-ws8bf" event={"ID":"d2302ab5-3f22-4f2e-804f-a9e1109446a6","Type":"ContainerStarted","Data":"d901eeba1e998f9a7d19f622be646399174c952260109f9a9fae3d653d7e652a"} Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.236149 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-ws8bf" event={"ID":"d2302ab5-3f22-4f2e-804f-a9e1109446a6","Type":"ContainerStarted","Data":"7d0045e2e784545e1752257ba2afd4820b1986ab34f8d4d3de7b50b29704a8fd"} Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.236158 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-ws8bf" event={"ID":"d2302ab5-3f22-4f2e-804f-a9e1109446a6","Type":"ContainerStarted","Data":"054a44c2eb52b525fe52381ca1659cf6994abe9c4b8ca2c1794ac72233ce0dfe"} Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.236245 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.255378 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-ws8bf" podStartSLOduration=2.255357187 podStartE2EDuration="2.255357187s" podCreationTimestamp="2025-12-02 10:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:33:41.248795887 +0000 UTC m=+874.578934747" watchObservedRunningTime="2025-12-02 10:33:41.255357187 +0000 UTC m=+874.585496047" Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.397549 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.403432 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c42448eb-c465-4172-9830-37000ff7e881-memberlist\") pod \"speaker-qgdwn\" (UID: \"c42448eb-c465-4172-9830-37000ff7e881\") " pod="metallb-system/speaker-qgdwn" Dec 02 10:33:41 crc kubenswrapper[4679]: I1202 10:33:41.584567 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qgdwn" Dec 02 10:33:42 crc kubenswrapper[4679]: I1202 10:33:42.244810 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qgdwn" event={"ID":"c42448eb-c465-4172-9830-37000ff7e881","Type":"ContainerStarted","Data":"eb6ef12cec41095f8d8ad1dcfb174c4d50b612ec84e2828d1ee9198ea13cfb7f"} Dec 02 10:33:42 crc kubenswrapper[4679]: I1202 10:33:42.245223 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qgdwn" event={"ID":"c42448eb-c465-4172-9830-37000ff7e881","Type":"ContainerStarted","Data":"bb38df90c2dcdeacb94859abb1642c379a8ba8b119b09523a1f05addd892bf76"} Dec 02 10:33:42 crc kubenswrapper[4679]: I1202 10:33:42.245238 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qgdwn" event={"ID":"c42448eb-c465-4172-9830-37000ff7e881","Type":"ContainerStarted","Data":"ec476581538c3912689b0a02166d47cd343aa3a42b6c45fb13b5f246813739a6"} Dec 02 10:33:42 crc kubenswrapper[4679]: I1202 10:33:42.245368 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qgdwn" Dec 02 10:33:42 crc kubenswrapper[4679]: I1202 10:33:42.277528 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qgdwn" podStartSLOduration=3.27750943 podStartE2EDuration="3.27750943s" podCreationTimestamp="2025-12-02 10:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:33:42.276794609 +0000 UTC m=+875.606933469" watchObservedRunningTime="2025-12-02 10:33:42.27750943 +0000 UTC m=+875.607648300" Dec 02 10:33:48 crc kubenswrapper[4679]: I1202 10:33:48.282811 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" event={"ID":"280f2b1e-4e5a-481d-ad2f-4536d832697a","Type":"ContainerStarted","Data":"5432c177f660fbae3d4902431beab501b3c42dcdcb93824cfba88d76e2bd9461"} Dec 02 10:33:48 crc kubenswrapper[4679]: I1202 10:33:48.283460 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:33:48 crc kubenswrapper[4679]: I1202 10:33:48.286899 4679 generic.go:334] "Generic (PLEG): container finished" podID="c5c9e3af-4021-4100-8a4f-3523f5da7bf9" containerID="f800af873a498b632ffdd09c45dfa43a9cb4a1125a2ab8c3a59b71bd218588a9" exitCode=0 Dec 02 10:33:48 crc kubenswrapper[4679]: I1202 10:33:48.286949 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerDied","Data":"f800af873a498b632ffdd09c45dfa43a9cb4a1125a2ab8c3a59b71bd218588a9"} Dec 02 10:33:48 crc kubenswrapper[4679]: I1202 10:33:48.328100 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" podStartSLOduration=2.673102924 podStartE2EDuration="9.3280769s" podCreationTimestamp="2025-12-02 10:33:39 +0000 UTC" firstStartedPulling="2025-12-02 10:33:40.672947139 +0000 UTC m=+874.003085999" lastFinishedPulling="2025-12-02 10:33:47.327921115 +0000 UTC m=+880.658059975" observedRunningTime="2025-12-02 10:33:48.323276871 +0000 UTC m=+881.653415731" watchObservedRunningTime="2025-12-02 10:33:48.3280769 +0000 UTC m=+881.658215760" Dec 02 10:33:49 crc kubenswrapper[4679]: I1202 10:33:49.295512 4679 generic.go:334] "Generic (PLEG): container finished" podID="c5c9e3af-4021-4100-8a4f-3523f5da7bf9" containerID="ac490a4405f3b344415abfa5d5030b8ca65df3e980c0646c1ab78d477a503107" exitCode=0 Dec 02 10:33:49 crc kubenswrapper[4679]: I1202 10:33:49.295575 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerDied","Data":"ac490a4405f3b344415abfa5d5030b8ca65df3e980c0646c1ab78d477a503107"} Dec 02 10:33:50 crc kubenswrapper[4679]: I1202 10:33:50.134392 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-ws8bf" Dec 02 10:33:50 crc kubenswrapper[4679]: I1202 10:33:50.304548 4679 generic.go:334] "Generic (PLEG): container finished" podID="c5c9e3af-4021-4100-8a4f-3523f5da7bf9" containerID="80117270d7474db557aefa304f26bd61c8e29e390fe81523a60413b60f9089ed" exitCode=0 Dec 02 10:33:50 crc kubenswrapper[4679]: I1202 10:33:50.304587 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerDied","Data":"80117270d7474db557aefa304f26bd61c8e29e390fe81523a60413b60f9089ed"} Dec 02 10:33:51 crc kubenswrapper[4679]: I1202 10:33:51.316329 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"0d9fc37129e1c9fcb502a70256d90edc9f9c7ea97d5ca819fcb16592d796bcfe"} Dec 02 10:33:51 crc kubenswrapper[4679]: I1202 10:33:51.589649 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qgdwn" Dec 02 10:33:52 crc kubenswrapper[4679]: I1202 10:33:52.327800 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"6c5d94d816cd98d133083ab54808c51e7513fed3b55e1d58b1974a9996405682"} Dec 02 10:33:52 crc kubenswrapper[4679]: I1202 10:33:52.327841 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"bf091943e81b87bbeeec141c5f205e4b3688da98c1d8818291c7f5897c5db07e"} Dec 02 10:33:52 crc kubenswrapper[4679]: I1202 10:33:52.327850 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"1f9f91db2de71dd13485618078ccc4069cdc923d83374f74179b6dcc1436f0fc"} Dec 02 10:33:52 crc kubenswrapper[4679]: I1202 10:33:52.327858 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"450ae1a64d668230f0fc4ae9791b0d6a110f7fc4ed18fa88aa89004af7d81b4b"} Dec 02 10:33:53 crc kubenswrapper[4679]: I1202 10:33:53.339098 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-qx64n" event={"ID":"c5c9e3af-4021-4100-8a4f-3523f5da7bf9","Type":"ContainerStarted","Data":"c722d6887293ab1ae1a7b73dc4560247e760f05c89f8f9b5885aeae82bc1447e"} Dec 02 10:33:53 crc kubenswrapper[4679]: I1202 10:33:53.339420 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:53 crc kubenswrapper[4679]: I1202 10:33:53.362675 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-qx64n" podStartSLOduration=7.070509304 podStartE2EDuration="14.362659507s" podCreationTimestamp="2025-12-02 10:33:39 +0000 UTC" firstStartedPulling="2025-12-02 10:33:40.019233743 +0000 UTC m=+873.349372603" lastFinishedPulling="2025-12-02 10:33:47.311383926 +0000 UTC m=+880.641522806" observedRunningTime="2025-12-02 10:33:53.359458534 +0000 UTC m=+886.689597444" watchObservedRunningTime="2025-12-02 10:33:53.362659507 +0000 UTC m=+886.692798367" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.591328 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pw7wk"] Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.592030 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.596329 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vbdb6" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.596542 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.596546 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.601636 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pw7wk"] Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.666699 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nzk2\" (UniqueName: \"kubernetes.io/projected/1e0d196d-2d14-4cc0-8f48-07769689e640-kube-api-access-8nzk2\") pod \"openstack-operator-index-pw7wk\" (UID: \"1e0d196d-2d14-4cc0-8f48-07769689e640\") " pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.767371 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nzk2\" (UniqueName: \"kubernetes.io/projected/1e0d196d-2d14-4cc0-8f48-07769689e640-kube-api-access-8nzk2\") pod \"openstack-operator-index-pw7wk\" (UID: \"1e0d196d-2d14-4cc0-8f48-07769689e640\") " pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.786158 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nzk2\" (UniqueName: \"kubernetes.io/projected/1e0d196d-2d14-4cc0-8f48-07769689e640-kube-api-access-8nzk2\") pod \"openstack-operator-index-pw7wk\" (UID: \"1e0d196d-2d14-4cc0-8f48-07769689e640\") " pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.873833 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.919849 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:33:54 crc kubenswrapper[4679]: I1202 10:33:54.948166 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-qx64n" Dec 02 10:33:55 crc kubenswrapper[4679]: I1202 10:33:55.155774 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pw7wk"] Dec 02 10:33:55 crc kubenswrapper[4679]: I1202 10:33:55.352057 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pw7wk" event={"ID":"1e0d196d-2d14-4cc0-8f48-07769689e640","Type":"ContainerStarted","Data":"58f581d22a0be5ef856105351d728a5e03974f6e771b47e7aeaaad7eebfbe371"} Dec 02 10:33:57 crc kubenswrapper[4679]: I1202 10:33:57.963219 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pw7wk"] Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.575729 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wh4rf"] Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.576511 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.580674 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wh4rf"] Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.720677 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfslv\" (UniqueName: \"kubernetes.io/projected/7fa3fa68-e237-45d4-b1b4-7422e84f46ef-kube-api-access-pfslv\") pod \"openstack-operator-index-wh4rf\" (UID: \"7fa3fa68-e237-45d4-b1b4-7422e84f46ef\") " pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.821794 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfslv\" (UniqueName: \"kubernetes.io/projected/7fa3fa68-e237-45d4-b1b4-7422e84f46ef-kube-api-access-pfslv\") pod \"openstack-operator-index-wh4rf\" (UID: \"7fa3fa68-e237-45d4-b1b4-7422e84f46ef\") " pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.839547 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfslv\" (UniqueName: \"kubernetes.io/projected/7fa3fa68-e237-45d4-b1b4-7422e84f46ef-kube-api-access-pfslv\") pod \"openstack-operator-index-wh4rf\" (UID: \"7fa3fa68-e237-45d4-b1b4-7422e84f46ef\") " pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:33:58 crc kubenswrapper[4679]: I1202 10:33:58.911532 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.380299 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pw7wk" event={"ID":"1e0d196d-2d14-4cc0-8f48-07769689e640","Type":"ContainerStarted","Data":"b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe"} Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.380447 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-pw7wk" podUID="1e0d196d-2d14-4cc0-8f48-07769689e640" containerName="registry-server" containerID="cri-o://b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe" gracePeriod=2 Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.401468 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pw7wk" podStartSLOduration=1.386479079 podStartE2EDuration="5.401442095s" podCreationTimestamp="2025-12-02 10:33:54 +0000 UTC" firstStartedPulling="2025-12-02 10:33:55.165172016 +0000 UTC m=+888.495310876" lastFinishedPulling="2025-12-02 10:33:59.180135032 +0000 UTC m=+892.510273892" observedRunningTime="2025-12-02 10:33:59.394613668 +0000 UTC m=+892.724752528" watchObservedRunningTime="2025-12-02 10:33:59.401442095 +0000 UTC m=+892.731580995" Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.563913 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wh4rf"] Dec 02 10:33:59 crc kubenswrapper[4679]: W1202 10:33:59.564273 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fa3fa68_e237_45d4_b1b4_7422e84f46ef.slice/crio-c2ef7711a1a612a0da576a99fd7428d25b4915d8cf2642ae6172d458e15e3cab WatchSource:0}: Error finding container c2ef7711a1a612a0da576a99fd7428d25b4915d8cf2642ae6172d458e15e3cab: Status 404 returned error can't find the container with id c2ef7711a1a612a0da576a99fd7428d25b4915d8cf2642ae6172d458e15e3cab Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.783655 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.938221 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nzk2\" (UniqueName: \"kubernetes.io/projected/1e0d196d-2d14-4cc0-8f48-07769689e640-kube-api-access-8nzk2\") pod \"1e0d196d-2d14-4cc0-8f48-07769689e640\" (UID: \"1e0d196d-2d14-4cc0-8f48-07769689e640\") " Dec 02 10:33:59 crc kubenswrapper[4679]: I1202 10:33:59.943454 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e0d196d-2d14-4cc0-8f48-07769689e640-kube-api-access-8nzk2" (OuterVolumeSpecName: "kube-api-access-8nzk2") pod "1e0d196d-2d14-4cc0-8f48-07769689e640" (UID: "1e0d196d-2d14-4cc0-8f48-07769689e640"). InnerVolumeSpecName "kube-api-access-8nzk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.040345 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nzk2\" (UniqueName: \"kubernetes.io/projected/1e0d196d-2d14-4cc0-8f48-07769689e640-kube-api-access-8nzk2\") on node \"crc\" DevicePath \"\"" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.390531 4679 generic.go:334] "Generic (PLEG): container finished" podID="1e0d196d-2d14-4cc0-8f48-07769689e640" containerID="b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe" exitCode=0 Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.390626 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pw7wk" event={"ID":"1e0d196d-2d14-4cc0-8f48-07769689e640","Type":"ContainerDied","Data":"b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe"} Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.390656 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pw7wk" event={"ID":"1e0d196d-2d14-4cc0-8f48-07769689e640","Type":"ContainerDied","Data":"58f581d22a0be5ef856105351d728a5e03974f6e771b47e7aeaaad7eebfbe371"} Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.390674 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pw7wk" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.390682 4679 scope.go:117] "RemoveContainer" containerID="b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.392903 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wh4rf" event={"ID":"7fa3fa68-e237-45d4-b1b4-7422e84f46ef","Type":"ContainerStarted","Data":"8e90194faca0daec27913ba2870d57843fce7e9fc872453854427e1ed7c63eb7"} Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.392962 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wh4rf" event={"ID":"7fa3fa68-e237-45d4-b1b4-7422e84f46ef","Type":"ContainerStarted","Data":"c2ef7711a1a612a0da576a99fd7428d25b4915d8cf2642ae6172d458e15e3cab"} Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.416129 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wh4rf" podStartSLOduration=2.357678158 podStartE2EDuration="2.416106842s" podCreationTimestamp="2025-12-02 10:33:58 +0000 UTC" firstStartedPulling="2025-12-02 10:33:59.570925937 +0000 UTC m=+892.901064817" lastFinishedPulling="2025-12-02 10:33:59.629354591 +0000 UTC m=+892.959493501" observedRunningTime="2025-12-02 10:34:00.413699952 +0000 UTC m=+893.743838812" watchObservedRunningTime="2025-12-02 10:34:00.416106842 +0000 UTC m=+893.746245702" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.428664 4679 scope.go:117] "RemoveContainer" containerID="b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe" Dec 02 10:34:00 crc kubenswrapper[4679]: E1202 10:34:00.429102 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe\": container with ID starting with b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe not found: ID does not exist" containerID="b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.429154 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe"} err="failed to get container status \"b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe\": rpc error: code = NotFound desc = could not find container \"b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe\": container with ID starting with b7e64c8370e76a3a8c23ebae859c670ff5bf214a5e6a7883548174c9b51320fe not found: ID does not exist" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.430249 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pw7wk"] Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.435473 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-pw7wk"] Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.468175 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-zcqd5" Dec 02 10:34:00 crc kubenswrapper[4679]: I1202 10:34:00.915282 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e0d196d-2d14-4cc0-8f48-07769689e640" path="/var/lib/kubelet/pods/1e0d196d-2d14-4cc0-8f48-07769689e640/volumes" Dec 02 10:34:08 crc kubenswrapper[4679]: I1202 10:34:08.918776 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:34:08 crc kubenswrapper[4679]: I1202 10:34:08.919511 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:34:08 crc kubenswrapper[4679]: I1202 10:34:08.943690 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:34:09 crc kubenswrapper[4679]: I1202 10:34:09.480650 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-wh4rf" Dec 02 10:34:09 crc kubenswrapper[4679]: I1202 10:34:09.878161 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-qx64n" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.347716 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44"] Dec 02 10:34:17 crc kubenswrapper[4679]: E1202 10:34:17.348655 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e0d196d-2d14-4cc0-8f48-07769689e640" containerName="registry-server" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.348672 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e0d196d-2d14-4cc0-8f48-07769689e640" containerName="registry-server" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.348796 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e0d196d-2d14-4cc0-8f48-07769689e640" containerName="registry-server" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.349602 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.353558 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44"] Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.353950 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2gdzz" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.379648 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-bundle\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.379856 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-util\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.379931 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8d4\" (UniqueName: \"kubernetes.io/projected/f5705dac-dfbc-4996-b0b2-177ce907b32a-kube-api-access-jk8d4\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.482056 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-util\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.482182 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8d4\" (UniqueName: \"kubernetes.io/projected/f5705dac-dfbc-4996-b0b2-177ce907b32a-kube-api-access-jk8d4\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.482338 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-bundle\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.482912 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-bundle\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.483036 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-util\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.507224 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8d4\" (UniqueName: \"kubernetes.io/projected/f5705dac-dfbc-4996-b0b2-177ce907b32a-kube-api-access-jk8d4\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:17 crc kubenswrapper[4679]: I1202 10:34:17.683163 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:18 crc kubenswrapper[4679]: I1202 10:34:18.126613 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44"] Dec 02 10:34:18 crc kubenswrapper[4679]: I1202 10:34:18.534702 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" event={"ID":"f5705dac-dfbc-4996-b0b2-177ce907b32a","Type":"ContainerStarted","Data":"98373c6c17f1e6cecc41f842d9f2770169fe8a5b22b50b08eb1c9efae16b7d85"} Dec 02 10:34:18 crc kubenswrapper[4679]: I1202 10:34:18.534759 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" event={"ID":"f5705dac-dfbc-4996-b0b2-177ce907b32a","Type":"ContainerStarted","Data":"262b5e23044de316154db89dc1e150c91d7682093fba5ac1864ef57334094d37"} Dec 02 10:34:19 crc kubenswrapper[4679]: I1202 10:34:19.542163 4679 generic.go:334] "Generic (PLEG): container finished" podID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerID="98373c6c17f1e6cecc41f842d9f2770169fe8a5b22b50b08eb1c9efae16b7d85" exitCode=0 Dec 02 10:34:19 crc kubenswrapper[4679]: I1202 10:34:19.542205 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" event={"ID":"f5705dac-dfbc-4996-b0b2-177ce907b32a","Type":"ContainerDied","Data":"98373c6c17f1e6cecc41f842d9f2770169fe8a5b22b50b08eb1c9efae16b7d85"} Dec 02 10:34:20 crc kubenswrapper[4679]: I1202 10:34:20.550986 4679 generic.go:334] "Generic (PLEG): container finished" podID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerID="1a696c0d1947387733ed7b426505e9e61f19fe1a5779e0ea689e2cbf2a0ebb16" exitCode=0 Dec 02 10:34:20 crc kubenswrapper[4679]: I1202 10:34:20.551050 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" event={"ID":"f5705dac-dfbc-4996-b0b2-177ce907b32a","Type":"ContainerDied","Data":"1a696c0d1947387733ed7b426505e9e61f19fe1a5779e0ea689e2cbf2a0ebb16"} Dec 02 10:34:21 crc kubenswrapper[4679]: I1202 10:34:21.561550 4679 generic.go:334] "Generic (PLEG): container finished" podID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerID="905fadb5e2351a9b8fbd96027bbb6554fc0673d5969b2553b8b2411b63a66774" exitCode=0 Dec 02 10:34:21 crc kubenswrapper[4679]: I1202 10:34:21.561628 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" event={"ID":"f5705dac-dfbc-4996-b0b2-177ce907b32a","Type":"ContainerDied","Data":"905fadb5e2351a9b8fbd96027bbb6554fc0673d5969b2553b8b2411b63a66774"} Dec 02 10:34:22 crc kubenswrapper[4679]: I1202 10:34:22.878561 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.054868 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-bundle\") pod \"f5705dac-dfbc-4996-b0b2-177ce907b32a\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.054952 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-util\") pod \"f5705dac-dfbc-4996-b0b2-177ce907b32a\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.054995 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk8d4\" (UniqueName: \"kubernetes.io/projected/f5705dac-dfbc-4996-b0b2-177ce907b32a-kube-api-access-jk8d4\") pod \"f5705dac-dfbc-4996-b0b2-177ce907b32a\" (UID: \"f5705dac-dfbc-4996-b0b2-177ce907b32a\") " Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.055860 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-bundle" (OuterVolumeSpecName: "bundle") pod "f5705dac-dfbc-4996-b0b2-177ce907b32a" (UID: "f5705dac-dfbc-4996-b0b2-177ce907b32a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.063809 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5705dac-dfbc-4996-b0b2-177ce907b32a-kube-api-access-jk8d4" (OuterVolumeSpecName: "kube-api-access-jk8d4") pod "f5705dac-dfbc-4996-b0b2-177ce907b32a" (UID: "f5705dac-dfbc-4996-b0b2-177ce907b32a"). InnerVolumeSpecName "kube-api-access-jk8d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.068446 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-util" (OuterVolumeSpecName: "util") pod "f5705dac-dfbc-4996-b0b2-177ce907b32a" (UID: "f5705dac-dfbc-4996-b0b2-177ce907b32a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.156038 4679 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.156078 4679 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5705dac-dfbc-4996-b0b2-177ce907b32a-util\") on node \"crc\" DevicePath \"\"" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.156086 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk8d4\" (UniqueName: \"kubernetes.io/projected/f5705dac-dfbc-4996-b0b2-177ce907b32a-kube-api-access-jk8d4\") on node \"crc\" DevicePath \"\"" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.581361 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" event={"ID":"f5705dac-dfbc-4996-b0b2-177ce907b32a","Type":"ContainerDied","Data":"262b5e23044de316154db89dc1e150c91d7682093fba5ac1864ef57334094d37"} Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.581410 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="262b5e23044de316154db89dc1e150c91d7682093fba5ac1864ef57334094d37" Dec 02 10:34:23 crc kubenswrapper[4679]: I1202 10:34:23.581439 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.683640 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld"] Dec 02 10:34:29 crc kubenswrapper[4679]: E1202 10:34:29.684501 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="pull" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.684517 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="pull" Dec 02 10:34:29 crc kubenswrapper[4679]: E1202 10:34:29.684533 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="extract" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.684540 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="extract" Dec 02 10:34:29 crc kubenswrapper[4679]: E1202 10:34:29.684556 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="util" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.684564 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="util" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.684681 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5705dac-dfbc-4996-b0b2-177ce907b32a" containerName="extract" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.685130 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.687390 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-sct2d" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.705660 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld"] Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.819967 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44xxb\" (UniqueName: \"kubernetes.io/projected/7b731918-9630-4abe-a4ce-ce8a65167295-kube-api-access-44xxb\") pod \"openstack-operator-controller-operator-655b495c79-ktwld\" (UID: \"7b731918-9630-4abe-a4ce-ce8a65167295\") " pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.920997 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44xxb\" (UniqueName: \"kubernetes.io/projected/7b731918-9630-4abe-a4ce-ce8a65167295-kube-api-access-44xxb\") pod \"openstack-operator-controller-operator-655b495c79-ktwld\" (UID: \"7b731918-9630-4abe-a4ce-ce8a65167295\") " pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:29 crc kubenswrapper[4679]: I1202 10:34:29.953465 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44xxb\" (UniqueName: \"kubernetes.io/projected/7b731918-9630-4abe-a4ce-ce8a65167295-kube-api-access-44xxb\") pod \"openstack-operator-controller-operator-655b495c79-ktwld\" (UID: \"7b731918-9630-4abe-a4ce-ce8a65167295\") " pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:30 crc kubenswrapper[4679]: I1202 10:34:30.004871 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:30 crc kubenswrapper[4679]: I1202 10:34:30.258699 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld"] Dec 02 10:34:30 crc kubenswrapper[4679]: W1202 10:34:30.271458 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b731918_9630_4abe_a4ce_ce8a65167295.slice/crio-7d526a1981923755a4d666e0970e36b62f525ee293c0c0d233a9f0721d54669e WatchSource:0}: Error finding container 7d526a1981923755a4d666e0970e36b62f525ee293c0c0d233a9f0721d54669e: Status 404 returned error can't find the container with id 7d526a1981923755a4d666e0970e36b62f525ee293c0c0d233a9f0721d54669e Dec 02 10:34:30 crc kubenswrapper[4679]: I1202 10:34:30.644279 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" event={"ID":"7b731918-9630-4abe-a4ce-ce8a65167295","Type":"ContainerStarted","Data":"7d526a1981923755a4d666e0970e36b62f525ee293c0c0d233a9f0721d54669e"} Dec 02 10:34:34 crc kubenswrapper[4679]: I1202 10:34:34.671374 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" event={"ID":"7b731918-9630-4abe-a4ce-ce8a65167295","Type":"ContainerStarted","Data":"a825ec13abad2fac3e88e1768be524edce569953884449e1899160ec9312d1c8"} Dec 02 10:34:34 crc kubenswrapper[4679]: I1202 10:34:34.672107 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:40 crc kubenswrapper[4679]: I1202 10:34:40.008029 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" Dec 02 10:34:40 crc kubenswrapper[4679]: I1202 10:34:40.060989 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-655b495c79-ktwld" podStartSLOduration=7.273917083 podStartE2EDuration="11.060956355s" podCreationTimestamp="2025-12-02 10:34:29 +0000 UTC" firstStartedPulling="2025-12-02 10:34:30.273974758 +0000 UTC m=+923.604113618" lastFinishedPulling="2025-12-02 10:34:34.06101403 +0000 UTC m=+927.391152890" observedRunningTime="2025-12-02 10:34:34.713620134 +0000 UTC m=+928.043759004" watchObservedRunningTime="2025-12-02 10:34:40.060956355 +0000 UTC m=+933.391095245" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.449200 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vnrzb"] Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.450777 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.474464 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnrzb"] Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.584983 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-catalog-content\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.585036 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgnfd\" (UniqueName: \"kubernetes.io/projected/d521c0f0-970a-4d99-b361-5493fc29b6b6-kube-api-access-vgnfd\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.585068 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-utilities\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.686499 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-utilities\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.687057 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-utilities\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.687279 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-catalog-content\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.687406 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgnfd\" (UniqueName: \"kubernetes.io/projected/d521c0f0-970a-4d99-b361-5493fc29b6b6-kube-api-access-vgnfd\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.687719 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-catalog-content\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.705859 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgnfd\" (UniqueName: \"kubernetes.io/projected/d521c0f0-970a-4d99-b361-5493fc29b6b6-kube-api-access-vgnfd\") pod \"redhat-marketplace-vnrzb\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.773246 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:49 crc kubenswrapper[4679]: I1202 10:34:49.985595 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnrzb"] Dec 02 10:34:50 crc kubenswrapper[4679]: I1202 10:34:50.786994 4679 generic.go:334] "Generic (PLEG): container finished" podID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerID="98a13cf5aece920032cbde050bf394c3198e45df19f1af05a32664f3d49e9a1d" exitCode=0 Dec 02 10:34:50 crc kubenswrapper[4679]: I1202 10:34:50.787139 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnrzb" event={"ID":"d521c0f0-970a-4d99-b361-5493fc29b6b6","Type":"ContainerDied","Data":"98a13cf5aece920032cbde050bf394c3198e45df19f1af05a32664f3d49e9a1d"} Dec 02 10:34:50 crc kubenswrapper[4679]: I1202 10:34:50.787469 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnrzb" event={"ID":"d521c0f0-970a-4d99-b361-5493fc29b6b6","Type":"ContainerStarted","Data":"c8d2923567b9524658084f2a9c307371d3f72eff0af7cc0f8366b8cb874c3bd9"} Dec 02 10:34:52 crc kubenswrapper[4679]: I1202 10:34:52.800590 4679 generic.go:334] "Generic (PLEG): container finished" podID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerID="990006b744fed41d9eb51655e44d3f3de8b1f6e4598dffc204a73567be683bcb" exitCode=0 Dec 02 10:34:52 crc kubenswrapper[4679]: I1202 10:34:52.801081 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnrzb" event={"ID":"d521c0f0-970a-4d99-b361-5493fc29b6b6","Type":"ContainerDied","Data":"990006b744fed41d9eb51655e44d3f3de8b1f6e4598dffc204a73567be683bcb"} Dec 02 10:34:53 crc kubenswrapper[4679]: I1202 10:34:53.809077 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnrzb" event={"ID":"d521c0f0-970a-4d99-b361-5493fc29b6b6","Type":"ContainerStarted","Data":"fc92af4427a76a5f4d799fdffa7de411fa2edc1484c02ceede0c8704d3e87b15"} Dec 02 10:34:53 crc kubenswrapper[4679]: I1202 10:34:53.831099 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vnrzb" podStartSLOduration=2.358378686 podStartE2EDuration="4.831077253s" podCreationTimestamp="2025-12-02 10:34:49 +0000 UTC" firstStartedPulling="2025-12-02 10:34:50.788577516 +0000 UTC m=+944.118716376" lastFinishedPulling="2025-12-02 10:34:53.261276083 +0000 UTC m=+946.591414943" observedRunningTime="2025-12-02 10:34:53.825450431 +0000 UTC m=+947.155589291" watchObservedRunningTime="2025-12-02 10:34:53.831077253 +0000 UTC m=+947.161216123" Dec 02 10:34:59 crc kubenswrapper[4679]: I1202 10:34:59.773750 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:59 crc kubenswrapper[4679]: I1202 10:34:59.774943 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:59 crc kubenswrapper[4679]: I1202 10:34:59.827135 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:34:59 crc kubenswrapper[4679]: I1202 10:34:59.896812 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:35:02 crc kubenswrapper[4679]: I1202 10:35:02.240012 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnrzb"] Dec 02 10:35:02 crc kubenswrapper[4679]: I1202 10:35:02.240992 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vnrzb" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="registry-server" containerID="cri-o://fc92af4427a76a5f4d799fdffa7de411fa2edc1484c02ceede0c8704d3e87b15" gracePeriod=2 Dec 02 10:35:03 crc kubenswrapper[4679]: I1202 10:35:03.876603 4679 generic.go:334] "Generic (PLEG): container finished" podID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerID="fc92af4427a76a5f4d799fdffa7de411fa2edc1484c02ceede0c8704d3e87b15" exitCode=0 Dec 02 10:35:03 crc kubenswrapper[4679]: I1202 10:35:03.876673 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnrzb" event={"ID":"d521c0f0-970a-4d99-b361-5493fc29b6b6","Type":"ContainerDied","Data":"fc92af4427a76a5f4d799fdffa7de411fa2edc1484c02ceede0c8704d3e87b15"} Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.337366 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.414390 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-utilities\") pod \"d521c0f0-970a-4d99-b361-5493fc29b6b6\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.414446 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-catalog-content\") pod \"d521c0f0-970a-4d99-b361-5493fc29b6b6\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.414522 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgnfd\" (UniqueName: \"kubernetes.io/projected/d521c0f0-970a-4d99-b361-5493fc29b6b6-kube-api-access-vgnfd\") pod \"d521c0f0-970a-4d99-b361-5493fc29b6b6\" (UID: \"d521c0f0-970a-4d99-b361-5493fc29b6b6\") " Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.416649 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-utilities" (OuterVolumeSpecName: "utilities") pod "d521c0f0-970a-4d99-b361-5493fc29b6b6" (UID: "d521c0f0-970a-4d99-b361-5493fc29b6b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.432041 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d521c0f0-970a-4d99-b361-5493fc29b6b6-kube-api-access-vgnfd" (OuterVolumeSpecName: "kube-api-access-vgnfd") pod "d521c0f0-970a-4d99-b361-5493fc29b6b6" (UID: "d521c0f0-970a-4d99-b361-5493fc29b6b6"). InnerVolumeSpecName "kube-api-access-vgnfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.434332 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d521c0f0-970a-4d99-b361-5493fc29b6b6" (UID: "d521c0f0-970a-4d99-b361-5493fc29b6b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.516224 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.516250 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d521c0f0-970a-4d99-b361-5493fc29b6b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.516260 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgnfd\" (UniqueName: \"kubernetes.io/projected/d521c0f0-970a-4d99-b361-5493fc29b6b6-kube-api-access-vgnfd\") on node \"crc\" DevicePath \"\"" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.889419 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnrzb" event={"ID":"d521c0f0-970a-4d99-b361-5493fc29b6b6","Type":"ContainerDied","Data":"c8d2923567b9524658084f2a9c307371d3f72eff0af7cc0f8366b8cb874c3bd9"} Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.889453 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnrzb" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.889764 4679 scope.go:117] "RemoveContainer" containerID="fc92af4427a76a5f4d799fdffa7de411fa2edc1484c02ceede0c8704d3e87b15" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.908922 4679 scope.go:117] "RemoveContainer" containerID="990006b744fed41d9eb51655e44d3f3de8b1f6e4598dffc204a73567be683bcb" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.917385 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnrzb"] Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.922449 4679 scope.go:117] "RemoveContainer" containerID="98a13cf5aece920032cbde050bf394c3198e45df19f1af05a32664f3d49e9a1d" Dec 02 10:35:05 crc kubenswrapper[4679]: I1202 10:35:05.928849 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnrzb"] Dec 02 10:35:06 crc kubenswrapper[4679]: I1202 10:35:06.916274 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" path="/var/lib/kubelet/pods/d521c0f0-970a-4d99-b361-5493fc29b6b6/volumes" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.617410 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d"] Dec 02 10:35:16 crc kubenswrapper[4679]: E1202 10:35:16.618103 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="extract-utilities" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.618115 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="extract-utilities" Dec 02 10:35:16 crc kubenswrapper[4679]: E1202 10:35:16.618135 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="registry-server" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.618143 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="registry-server" Dec 02 10:35:16 crc kubenswrapper[4679]: E1202 10:35:16.618152 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="extract-content" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.618160 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="extract-content" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.618259 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d521c0f0-970a-4d99-b361-5493fc29b6b6" containerName="registry-server" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.618833 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.620838 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-89rcp" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.627858 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.638413 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.639360 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.652836 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jm4xh" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.658255 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.659514 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.661182 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cndr8" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.669274 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.681106 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.702940 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.704067 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.709819 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pqqg7" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.724114 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.733622 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.734917 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.755533 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.757676 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.756479 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-d7t88" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.782317 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mns42\" (UniqueName: \"kubernetes.io/projected/7938c8f2-53f6-4143-997d-a44ea54c9fc2-kube-api-access-mns42\") pod \"cinder-operator-controller-manager-859b6ccc6-snss6\" (UID: \"7938c8f2-53f6-4143-997d-a44ea54c9fc2\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.782435 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx7gs\" (UniqueName: \"kubernetes.io/projected/48f2b621-5d9b-4213-8681-8107d1fc04a4-kube-api-access-dx7gs\") pod \"barbican-operator-controller-manager-7d9dfd778-jmj2d\" (UID: \"48f2b621-5d9b-4213-8681-8107d1fc04a4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.782476 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nxx5\" (UniqueName: \"kubernetes.io/projected/85ed4b5c-eda1-481a-84ae-b97d01a9b9eb-kube-api-access-2nxx5\") pod \"designate-operator-controller-manager-78b4bc895b-59k6f\" (UID: \"85ed4b5c-eda1-481a-84ae-b97d01a9b9eb\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.782623 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.783047 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-dphjd" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.828351 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.829676 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.834374 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.839931 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.840178 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jhvpc" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.852358 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.889471 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx7gs\" (UniqueName: \"kubernetes.io/projected/48f2b621-5d9b-4213-8681-8107d1fc04a4-kube-api-access-dx7gs\") pod \"barbican-operator-controller-manager-7d9dfd778-jmj2d\" (UID: \"48f2b621-5d9b-4213-8681-8107d1fc04a4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.889543 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nxx5\" (UniqueName: \"kubernetes.io/projected/85ed4b5c-eda1-481a-84ae-b97d01a9b9eb-kube-api-access-2nxx5\") pod \"designate-operator-controller-manager-78b4bc895b-59k6f\" (UID: \"85ed4b5c-eda1-481a-84ae-b97d01a9b9eb\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.889603 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkrjr\" (UniqueName: \"kubernetes.io/projected/80f86afb-0d79-4d1c-a0fd-b1d573d3c471-kube-api-access-mkrjr\") pod \"horizon-operator-controller-manager-68c6d99b8f-bnz2w\" (UID: \"80f86afb-0d79-4d1c-a0fd-b1d573d3c471\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.889654 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mns42\" (UniqueName: \"kubernetes.io/projected/7938c8f2-53f6-4143-997d-a44ea54c9fc2-kube-api-access-mns42\") pod \"cinder-operator-controller-manager-859b6ccc6-snss6\" (UID: \"7938c8f2-53f6-4143-997d-a44ea54c9fc2\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.889703 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhw6\" (UniqueName: \"kubernetes.io/projected/61bf694c-dbb6-442c-aaa0-3c13ff7efc97-kube-api-access-vmhw6\") pod \"glance-operator-controller-manager-77987cd8cd-442vr\" (UID: \"61bf694c-dbb6-442c-aaa0-3c13ff7efc97\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.889730 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4msg8\" (UniqueName: \"kubernetes.io/projected/f9cf4107-9e23-4246-a8b7-b3bbda0cc461-kube-api-access-4msg8\") pod \"heat-operator-controller-manager-5f64f6f8bb-jxm5w\" (UID: \"f9cf4107-9e23-4246-a8b7-b3bbda0cc461\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.956749 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mns42\" (UniqueName: \"kubernetes.io/projected/7938c8f2-53f6-4143-997d-a44ea54c9fc2-kube-api-access-mns42\") pod \"cinder-operator-controller-manager-859b6ccc6-snss6\" (UID: \"7938c8f2-53f6-4143-997d-a44ea54c9fc2\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.959041 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.961723 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd"] Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.999203 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:16 crc kubenswrapper[4679]: I1202 10:35:16.999892 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkrjr\" (UniqueName: \"kubernetes.io/projected/80f86afb-0d79-4d1c-a0fd-b1d573d3c471-kube-api-access-mkrjr\") pod \"horizon-operator-controller-manager-68c6d99b8f-bnz2w\" (UID: \"80f86afb-0d79-4d1c-a0fd-b1d573d3c471\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.000017 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2kf9\" (UniqueName: \"kubernetes.io/projected/a0088242-8397-41e7-96b8-d971f307eefc-kube-api-access-p2kf9\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.000111 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmhw6\" (UniqueName: \"kubernetes.io/projected/61bf694c-dbb6-442c-aaa0-3c13ff7efc97-kube-api-access-vmhw6\") pod \"glance-operator-controller-manager-77987cd8cd-442vr\" (UID: \"61bf694c-dbb6-442c-aaa0-3c13ff7efc97\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.000217 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4msg8\" (UniqueName: \"kubernetes.io/projected/f9cf4107-9e23-4246-a8b7-b3bbda0cc461-kube-api-access-4msg8\") pod \"heat-operator-controller-manager-5f64f6f8bb-jxm5w\" (UID: \"f9cf4107-9e23-4246-a8b7-b3bbda0cc461\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.005610 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.015105 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx7gs\" (UniqueName: \"kubernetes.io/projected/48f2b621-5d9b-4213-8681-8107d1fc04a4-kube-api-access-dx7gs\") pod \"barbican-operator-controller-manager-7d9dfd778-jmj2d\" (UID: \"48f2b621-5d9b-4213-8681-8107d1fc04a4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.027501 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nxx5\" (UniqueName: \"kubernetes.io/projected/85ed4b5c-eda1-481a-84ae-b97d01a9b9eb-kube-api-access-2nxx5\") pod \"designate-operator-controller-manager-78b4bc895b-59k6f\" (UID: \"85ed4b5c-eda1-481a-84ae-b97d01a9b9eb\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.037160 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.042485 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.045750 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-bckk2" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.049031 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.052332 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmhw6\" (UniqueName: \"kubernetes.io/projected/61bf694c-dbb6-442c-aaa0-3c13ff7efc97-kube-api-access-vmhw6\") pod \"glance-operator-controller-manager-77987cd8cd-442vr\" (UID: \"61bf694c-dbb6-442c-aaa0-3c13ff7efc97\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.070361 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rz5ql" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.086749 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.088186 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.092056 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.094778 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4msg8\" (UniqueName: \"kubernetes.io/projected/f9cf4107-9e23-4246-a8b7-b3bbda0cc461-kube-api-access-4msg8\") pod \"heat-operator-controller-manager-5f64f6f8bb-jxm5w\" (UID: \"f9cf4107-9e23-4246-a8b7-b3bbda0cc461\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.096538 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gmlgh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.100932 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpw4\" (UniqueName: \"kubernetes.io/projected/1cf6ea83-714b-4a9d-8a5e-708293a3e9e6-kube-api-access-fzpw4\") pod \"manila-operator-controller-manager-7c79b5df47-tp7zw\" (UID: \"1cf6ea83-714b-4a9d-8a5e-708293a3e9e6\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.100989 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.101027 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pdk5\" (UniqueName: \"kubernetes.io/projected/ebbb7257-823e-49a1-af4c-703b979de700-kube-api-access-8pdk5\") pod \"keystone-operator-controller-manager-7765d96ddf-sxgbw\" (UID: \"ebbb7257-823e-49a1-af4c-703b979de700\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.101058 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2kf9\" (UniqueName: \"kubernetes.io/projected/a0088242-8397-41e7-96b8-d971f307eefc-kube-api-access-p2kf9\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.101080 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvs52\" (UniqueName: \"kubernetes.io/projected/5ee03695-d4bb-4b77-84d1-84190ae307d6-kube-api-access-mvs52\") pod \"ironic-operator-controller-manager-6c548fd776-5zzmd\" (UID: \"5ee03695-d4bb-4b77-84d1-84190ae307d6\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:17 crc kubenswrapper[4679]: E1202 10:35:17.101248 4679 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:17 crc kubenswrapper[4679]: E1202 10:35:17.101325 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert podName:a0088242-8397-41e7-96b8-d971f307eefc nodeName:}" failed. No retries permitted until 2025-12-02 10:35:17.601280673 +0000 UTC m=+970.931419543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert") pod "infra-operator-controller-manager-57548d458d-ghgsb" (UID: "a0088242-8397-41e7-96b8-d971f307eefc") : secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.102025 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkrjr\" (UniqueName: \"kubernetes.io/projected/80f86afb-0d79-4d1c-a0fd-b1d573d3c471-kube-api-access-mkrjr\") pod \"horizon-operator-controller-manager-68c6d99b8f-bnz2w\" (UID: \"80f86afb-0d79-4d1c-a0fd-b1d573d3c471\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.108617 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.139121 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2kf9\" (UniqueName: \"kubernetes.io/projected/a0088242-8397-41e7-96b8-d971f307eefc-kube-api-access-p2kf9\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.141030 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.142031 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.148106 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-lhxwt" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.159413 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.160885 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.199874 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-628rp"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.201508 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.201871 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-ptjvr" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.208881 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pdk5\" (UniqueName: \"kubernetes.io/projected/ebbb7257-823e-49a1-af4c-703b979de700-kube-api-access-8pdk5\") pod \"keystone-operator-controller-manager-7765d96ddf-sxgbw\" (UID: \"ebbb7257-823e-49a1-af4c-703b979de700\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.208955 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvs52\" (UniqueName: \"kubernetes.io/projected/5ee03695-d4bb-4b77-84d1-84190ae307d6-kube-api-access-mvs52\") pod \"ironic-operator-controller-manager-6c548fd776-5zzmd\" (UID: \"5ee03695-d4bb-4b77-84d1-84190ae307d6\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.209000 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5sbx\" (UniqueName: \"kubernetes.io/projected/e334fbef-5aab-4cd6-83a1-084160e32d90-kube-api-access-f5sbx\") pod \"mariadb-operator-controller-manager-56bbcc9d85-w74h9\" (UID: \"e334fbef-5aab-4cd6-83a1-084160e32d90\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.220674 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-lcsgp" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.225066 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.225331 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpw4\" (UniqueName: \"kubernetes.io/projected/1cf6ea83-714b-4a9d-8a5e-708293a3e9e6-kube-api-access-fzpw4\") pod \"manila-operator-controller-manager-7c79b5df47-tp7zw\" (UID: \"1cf6ea83-714b-4a9d-8a5e-708293a3e9e6\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.239860 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.265404 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvs52\" (UniqueName: \"kubernetes.io/projected/5ee03695-d4bb-4b77-84d1-84190ae307d6-kube-api-access-mvs52\") pod \"ironic-operator-controller-manager-6c548fd776-5zzmd\" (UID: \"5ee03695-d4bb-4b77-84d1-84190ae307d6\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.270179 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pdk5\" (UniqueName: \"kubernetes.io/projected/ebbb7257-823e-49a1-af4c-703b979de700-kube-api-access-8pdk5\") pod \"keystone-operator-controller-manager-7765d96ddf-sxgbw\" (UID: \"ebbb7257-823e-49a1-af4c-703b979de700\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.270276 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.281362 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.293013 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-628rp"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.302948 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpw4\" (UniqueName: \"kubernetes.io/projected/1cf6ea83-714b-4a9d-8a5e-708293a3e9e6-kube-api-access-fzpw4\") pod \"manila-operator-controller-manager-7c79b5df47-tp7zw\" (UID: \"1cf6ea83-714b-4a9d-8a5e-708293a3e9e6\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.326997 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5sbx\" (UniqueName: \"kubernetes.io/projected/e334fbef-5aab-4cd6-83a1-084160e32d90-kube-api-access-f5sbx\") pod \"mariadb-operator-controller-manager-56bbcc9d85-w74h9\" (UID: \"e334fbef-5aab-4cd6-83a1-084160e32d90\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.327054 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6q7d\" (UniqueName: \"kubernetes.io/projected/e23de616-947b-4ff1-b05b-522a1c39cc7b-kube-api-access-q6q7d\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-48bsh\" (UID: \"e23de616-947b-4ff1-b05b-522a1c39cc7b\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.327117 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th2gd\" (UniqueName: \"kubernetes.io/projected/9dc9ab59-e368-452c-9ce2-e4d028fd35a1-kube-api-access-th2gd\") pod \"nova-operator-controller-manager-697bc559fc-628rp\" (UID: \"9dc9ab59-e368-452c-9ce2-e4d028fd35a1\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.333216 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.355353 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-nm552"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.356630 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.357641 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.363780 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-27c4z" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.380710 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.388619 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-nm552"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.413904 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5sbx\" (UniqueName: \"kubernetes.io/projected/e334fbef-5aab-4cd6-83a1-084160e32d90-kube-api-access-f5sbx\") pod \"mariadb-operator-controller-manager-56bbcc9d85-w74h9\" (UID: \"e334fbef-5aab-4cd6-83a1-084160e32d90\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.426839 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.429214 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th2gd\" (UniqueName: \"kubernetes.io/projected/9dc9ab59-e368-452c-9ce2-e4d028fd35a1-kube-api-access-th2gd\") pod \"nova-operator-controller-manager-697bc559fc-628rp\" (UID: \"9dc9ab59-e368-452c-9ce2-e4d028fd35a1\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.429323 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vccbj\" (UniqueName: \"kubernetes.io/projected/1b3b0e6e-eb2f-40fc-a77f-c8889816dee3-kube-api-access-vccbj\") pod \"octavia-operator-controller-manager-998648c74-nm552\" (UID: \"1b3b0e6e-eb2f-40fc-a77f-c8889816dee3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.429359 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6q7d\" (UniqueName: \"kubernetes.io/projected/e23de616-947b-4ff1-b05b-522a1c39cc7b-kube-api-access-q6q7d\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-48bsh\" (UID: \"e23de616-947b-4ff1-b05b-522a1c39cc7b\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.430408 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.442683 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.443981 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.444544 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.444598 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-jtc68" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.446333 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.451666 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-55bwf" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.471664 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.493998 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th2gd\" (UniqueName: \"kubernetes.io/projected/9dc9ab59-e368-452c-9ce2-e4d028fd35a1-kube-api-access-th2gd\") pod \"nova-operator-controller-manager-697bc559fc-628rp\" (UID: \"9dc9ab59-e368-452c-9ce2-e4d028fd35a1\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.497648 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.499045 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6q7d\" (UniqueName: \"kubernetes.io/projected/e23de616-947b-4ff1-b05b-522a1c39cc7b-kube-api-access-q6q7d\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-48bsh\" (UID: \"e23de616-947b-4ff1-b05b-522a1c39cc7b\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.510753 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-9flx2"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.511864 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.522064 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-k542g" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.533671 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8jnm\" (UniqueName: \"kubernetes.io/projected/3fa279c2-e6ef-4131-ab48-28306a24000e-kube-api-access-l8jnm\") pod \"ovn-operator-controller-manager-b6456fdb6-xfz4c\" (UID: \"3fa279c2-e6ef-4131-ab48-28306a24000e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.533761 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrhbm\" (UniqueName: \"kubernetes.io/projected/c26eed77-aa87-4b6b-9f81-aefe3e5ed786-kube-api-access-mrhbm\") pod \"placement-operator-controller-manager-78f8948974-9flx2\" (UID: \"c26eed77-aa87-4b6b-9f81-aefe3e5ed786\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.533816 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.533865 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vccbj\" (UniqueName: \"kubernetes.io/projected/1b3b0e6e-eb2f-40fc-a77f-c8889816dee3-kube-api-access-vccbj\") pod \"octavia-operator-controller-manager-998648c74-nm552\" (UID: \"1b3b0e6e-eb2f-40fc-a77f-c8889816dee3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.533904 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vnnl\" (UniqueName: \"kubernetes.io/projected/b255bbc9-0452-4445-9470-8bd7ad85257a-kube-api-access-7vnnl\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.541081 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.552125 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.568698 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.574604 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.574747 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.575154 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vccbj\" (UniqueName: \"kubernetes.io/projected/1b3b0e6e-eb2f-40fc-a77f-c8889816dee3-kube-api-access-vccbj\") pod \"octavia-operator-controller-manager-998648c74-nm552\" (UID: \"1b3b0e6e-eb2f-40fc-a77f-c8889816dee3\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.578711 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-98kqg" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.582388 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.595710 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.618591 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-9flx2"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.635942 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vnnl\" (UniqueName: \"kubernetes.io/projected/b255bbc9-0452-4445-9470-8bd7ad85257a-kube-api-access-7vnnl\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.635991 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8jnm\" (UniqueName: \"kubernetes.io/projected/3fa279c2-e6ef-4131-ab48-28306a24000e-kube-api-access-l8jnm\") pod \"ovn-operator-controller-manager-b6456fdb6-xfz4c\" (UID: \"3fa279c2-e6ef-4131-ab48-28306a24000e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.636021 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lvdc\" (UniqueName: \"kubernetes.io/projected/a877b082-6f7d-4f85-b290-356fe7c8bb5f-kube-api-access-5lvdc\") pod \"swift-operator-controller-manager-5f8c65bbfc-nw7fm\" (UID: \"a877b082-6f7d-4f85-b290-356fe7c8bb5f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.636049 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.636069 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrhbm\" (UniqueName: \"kubernetes.io/projected/c26eed77-aa87-4b6b-9f81-aefe3e5ed786-kube-api-access-mrhbm\") pod \"placement-operator-controller-manager-78f8948974-9flx2\" (UID: \"c26eed77-aa87-4b6b-9f81-aefe3e5ed786\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.636106 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:17 crc kubenswrapper[4679]: E1202 10:35:17.636233 4679 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:17 crc kubenswrapper[4679]: E1202 10:35:17.636277 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert podName:b255bbc9-0452-4445-9470-8bd7ad85257a nodeName:}" failed. No retries permitted until 2025-12-02 10:35:18.136263045 +0000 UTC m=+971.466401905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" (UID: "b255bbc9-0452-4445-9470-8bd7ad85257a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:17 crc kubenswrapper[4679]: E1202 10:35:17.636783 4679 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:17 crc kubenswrapper[4679]: E1202 10:35:17.636811 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert podName:a0088242-8397-41e7-96b8-d971f307eefc nodeName:}" failed. No retries permitted until 2025-12-02 10:35:18.636801381 +0000 UTC m=+971.966940241 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert") pod "infra-operator-controller-manager-57548d458d-ghgsb" (UID: "a0088242-8397-41e7-96b8-d971f307eefc") : secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.693782 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.710863 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vnnl\" (UniqueName: \"kubernetes.io/projected/b255bbc9-0452-4445-9470-8bd7ad85257a-kube-api-access-7vnnl\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.713799 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.714273 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.720832 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.723376 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-t7s2x" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.737286 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lvdc\" (UniqueName: \"kubernetes.io/projected/a877b082-6f7d-4f85-b290-356fe7c8bb5f-kube-api-access-5lvdc\") pod \"swift-operator-controller-manager-5f8c65bbfc-nw7fm\" (UID: \"a877b082-6f7d-4f85-b290-356fe7c8bb5f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.737365 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxn8b\" (UniqueName: \"kubernetes.io/projected/921e63bc-e66b-48d5-a22c-010c65a0e1b0-kube-api-access-gxn8b\") pod \"telemetry-operator-controller-manager-76cc84c6bb-ps9nb\" (UID: \"921e63bc-e66b-48d5-a22c-010c65a0e1b0\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.745496 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8jnm\" (UniqueName: \"kubernetes.io/projected/3fa279c2-e6ef-4131-ab48-28306a24000e-kube-api-access-l8jnm\") pod \"ovn-operator-controller-manager-b6456fdb6-xfz4c\" (UID: \"3fa279c2-e6ef-4131-ab48-28306a24000e\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.753245 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrhbm\" (UniqueName: \"kubernetes.io/projected/c26eed77-aa87-4b6b-9f81-aefe3e5ed786-kube-api-access-mrhbm\") pod \"placement-operator-controller-manager-78f8948974-9flx2\" (UID: \"c26eed77-aa87-4b6b-9f81-aefe3e5ed786\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.755264 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-b74hj"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.756347 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.760866 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rslzl" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.761499 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lvdc\" (UniqueName: \"kubernetes.io/projected/a877b082-6f7d-4f85-b290-356fe7c8bb5f-kube-api-access-5lvdc\") pod \"swift-operator-controller-manager-5f8c65bbfc-nw7fm\" (UID: \"a877b082-6f7d-4f85-b290-356fe7c8bb5f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.762390 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.771386 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-b74hj"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.783685 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.784744 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.790256 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.790502 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-4zrwl" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.797830 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.841424 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxn8b\" (UniqueName: \"kubernetes.io/projected/921e63bc-e66b-48d5-a22c-010c65a0e1b0-kube-api-access-gxn8b\") pod \"telemetry-operator-controller-manager-76cc84c6bb-ps9nb\" (UID: \"921e63bc-e66b-48d5-a22c-010c65a0e1b0\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.841693 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdtxx\" (UniqueName: \"kubernetes.io/projected/6ac36ac3-f2ba-40be-9544-9897a27ddaf4-kube-api-access-tdtxx\") pod \"test-operator-controller-manager-5854674fcc-b74hj\" (UID: \"6ac36ac3-f2ba-40be-9544-9897a27ddaf4\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.841771 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5xr5\" (UniqueName: \"kubernetes.io/projected/f48fce46-9ae8-4154-b67b-3edd56267209-kube-api-access-j5xr5\") pod \"watcher-operator-controller-manager-769dc69bc-7kl62\" (UID: \"f48fce46-9ae8-4154-b67b-3edd56267209\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.843883 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.845852 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.848168 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.848540 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.848808 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-m8wrc" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.860145 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.865480 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.876329 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxn8b\" (UniqueName: \"kubernetes.io/projected/921e63bc-e66b-48d5-a22c-010c65a0e1b0-kube-api-access-gxn8b\") pod \"telemetry-operator-controller-manager-76cc84c6bb-ps9nb\" (UID: \"921e63bc-e66b-48d5-a22c-010c65a0e1b0\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.879458 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.888350 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.890440 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.898115 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-rwrh4" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.899384 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn"] Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.927380 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.943243 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbzgx\" (UniqueName: \"kubernetes.io/projected/0a119e38-647a-42ab-823f-e5ecea4afbf0-kube-api-access-bbzgx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lxjdn\" (UID: \"0a119e38-647a-42ab-823f-e5ecea4afbf0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.943338 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.943408 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdtxx\" (UniqueName: \"kubernetes.io/projected/6ac36ac3-f2ba-40be-9544-9897a27ddaf4-kube-api-access-tdtxx\") pod \"test-operator-controller-manager-5854674fcc-b74hj\" (UID: \"6ac36ac3-f2ba-40be-9544-9897a27ddaf4\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.943467 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srmmz\" (UniqueName: \"kubernetes.io/projected/6edbf2ab-201d-4a7f-902a-7b74566b5a36-kube-api-access-srmmz\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.943489 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.943542 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5xr5\" (UniqueName: \"kubernetes.io/projected/f48fce46-9ae8-4154-b67b-3edd56267209-kube-api-access-j5xr5\") pod \"watcher-operator-controller-manager-769dc69bc-7kl62\" (UID: \"f48fce46-9ae8-4154-b67b-3edd56267209\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.960028 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdtxx\" (UniqueName: \"kubernetes.io/projected/6ac36ac3-f2ba-40be-9544-9897a27ddaf4-kube-api-access-tdtxx\") pod \"test-operator-controller-manager-5854674fcc-b74hj\" (UID: \"6ac36ac3-f2ba-40be-9544-9897a27ddaf4\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:17 crc kubenswrapper[4679]: I1202 10:35:17.964184 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5xr5\" (UniqueName: \"kubernetes.io/projected/f48fce46-9ae8-4154-b67b-3edd56267209-kube-api-access-j5xr5\") pod \"watcher-operator-controller-manager-769dc69bc-7kl62\" (UID: \"f48fce46-9ae8-4154-b67b-3edd56267209\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.042944 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d"] Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.045467 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srmmz\" (UniqueName: \"kubernetes.io/projected/6edbf2ab-201d-4a7f-902a-7b74566b5a36-kube-api-access-srmmz\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.045519 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.045595 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbzgx\" (UniqueName: \"kubernetes.io/projected/0a119e38-647a-42ab-823f-e5ecea4afbf0-kube-api-access-bbzgx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lxjdn\" (UID: \"0a119e38-647a-42ab-823f-e5ecea4afbf0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.045617 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.045738 4679 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.045766 4679 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.045586 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" event={"ID":"7938c8f2-53f6-4143-997d-a44ea54c9fc2","Type":"ContainerStarted","Data":"f55b2bdde39739024ec40aa895eb4397935721a1a8e8084f776dec595f1dffff"} Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.045805 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:18.54578327 +0000 UTC m=+971.875922210 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.045826 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:18.545817511 +0000 UTC m=+971.875956381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "metrics-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.053147 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.067617 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srmmz\" (UniqueName: \"kubernetes.io/projected/6edbf2ab-201d-4a7f-902a-7b74566b5a36-kube-api-access-srmmz\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.070489 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbzgx\" (UniqueName: \"kubernetes.io/projected/0a119e38-647a-42ab-823f-e5ecea4afbf0-kube-api-access-bbzgx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lxjdn\" (UID: \"0a119e38-647a-42ab-823f-e5ecea4afbf0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.098024 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.129489 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.147411 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.147542 4679 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.147590 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert podName:b255bbc9-0452-4445-9470-8bd7ad85257a nodeName:}" failed. No retries permitted until 2025-12-02 10:35:19.147575394 +0000 UTC m=+972.477714254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" (UID: "b255bbc9-0452-4445-9470-8bd7ad85257a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: W1202 10:35:18.185883 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f2b621_5d9b_4213_8681_8107d1fc04a4.slice/crio-cddd8cdc46c8d384ddb10c193bd4661d0a37a86a842e916f93eb46392c43e372 WatchSource:0}: Error finding container cddd8cdc46c8d384ddb10c193bd4661d0a37a86a842e916f93eb46392c43e372: Status 404 returned error can't find the container with id cddd8cdc46c8d384ddb10c193bd4661d0a37a86a842e916f93eb46392c43e372 Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.244613 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.554533 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.554748 4679 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.555022 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.555092 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:19.55506721 +0000 UTC m=+972.885206150 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.555221 4679 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.555365 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:19.555283826 +0000 UTC m=+972.885422686 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "metrics-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.605248 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f"] Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.641961 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr"] Dec 02 10:35:18 crc kubenswrapper[4679]: W1202 10:35:18.646084 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61bf694c_dbb6_442c_aaa0_3c13ff7efc97.slice/crio-8730506d85d4931ffa97590077f04a8df344b21f81740e0c157e183a3ba0b664 WatchSource:0}: Error finding container 8730506d85d4931ffa97590077f04a8df344b21f81740e0c157e183a3ba0b664: Status 404 returned error can't find the container with id 8730506d85d4931ffa97590077f04a8df344b21f81740e0c157e183a3ba0b664 Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.657426 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.657637 4679 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: E1202 10:35:18.657711 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert podName:a0088242-8397-41e7-96b8-d971f307eefc nodeName:}" failed. No retries permitted until 2025-12-02 10:35:20.657690358 +0000 UTC m=+973.987829288 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert") pod "infra-operator-controller-manager-57548d458d-ghgsb" (UID: "a0088242-8397-41e7-96b8-d971f307eefc") : secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.710910 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd"] Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.716022 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w"] Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.723950 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w"] Dec 02 10:35:18 crc kubenswrapper[4679]: I1202 10:35:18.737953 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh"] Dec 02 10:35:18 crc kubenswrapper[4679]: W1202 10:35:18.739852 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80f86afb_0d79_4d1c_a0fd_b1d573d3c471.slice/crio-5384092407ab1ddeed749195191134d9717f9cfca4bfc77de96d0aabebe84e3a WatchSource:0}: Error finding container 5384092407ab1ddeed749195191134d9717f9cfca4bfc77de96d0aabebe84e3a: Status 404 returned error can't find the container with id 5384092407ab1ddeed749195191134d9717f9cfca4bfc77de96d0aabebe84e3a Dec 02 10:35:18 crc kubenswrapper[4679]: W1202 10:35:18.740223 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9cf4107_9e23_4246_a8b7_b3bbda0cc461.slice/crio-5417c7bcfd2602304507de357620f62adc69ac958ac13cd999848a1490e779ea WatchSource:0}: Error finding container 5417c7bcfd2602304507de357620f62adc69ac958ac13cd999848a1490e779ea: Status 404 returned error can't find the container with id 5417c7bcfd2602304507de357620f62adc69ac958ac13cd999848a1490e779ea Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.053035 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" event={"ID":"f9cf4107-9e23-4246-a8b7-b3bbda0cc461","Type":"ContainerStarted","Data":"5417c7bcfd2602304507de357620f62adc69ac958ac13cd999848a1490e779ea"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.054197 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" event={"ID":"5ee03695-d4bb-4b77-84d1-84190ae307d6","Type":"ContainerStarted","Data":"0b53e0679d55a8d3ccd8bd2fbffce224c5af1bcbbedb1882f64bebcdbf502135"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.055362 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" event={"ID":"85ed4b5c-eda1-481a-84ae-b97d01a9b9eb","Type":"ContainerStarted","Data":"30ba99144e468a5a10e8c566d285d6f9fb3803dd2d21959385fd005b97178d05"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.056756 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" event={"ID":"61bf694c-dbb6-442c-aaa0-3c13ff7efc97","Type":"ContainerStarted","Data":"8730506d85d4931ffa97590077f04a8df344b21f81740e0c157e183a3ba0b664"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.057930 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" event={"ID":"80f86afb-0d79-4d1c-a0fd-b1d573d3c471","Type":"ContainerStarted","Data":"5384092407ab1ddeed749195191134d9717f9cfca4bfc77de96d0aabebe84e3a"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.059069 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" event={"ID":"48f2b621-5d9b-4213-8681-8107d1fc04a4","Type":"ContainerStarted","Data":"cddd8cdc46c8d384ddb10c193bd4661d0a37a86a842e916f93eb46392c43e372"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.060185 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" event={"ID":"e23de616-947b-4ff1-b05b-522a1c39cc7b","Type":"ContainerStarted","Data":"36fcba0d0664421ead887e6e50f26d370735b31c4fd6f499d44f05c0e8e57a7c"} Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.100018 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-9flx2"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.117442 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.126641 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.139653 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.149395 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-628rp"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.157940 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.166600 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-b74hj"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.169028 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.169216 4679 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.169285 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert podName:b255bbc9-0452-4445-9470-8bd7ad85257a nodeName:}" failed. No retries permitted until 2025-12-02 10:35:21.169251764 +0000 UTC m=+974.499390624 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" (UID: "b255bbc9-0452-4445-9470-8bd7ad85257a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.174987 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-nm552"] Dec 02 10:35:19 crc kubenswrapper[4679]: W1202 10:35:19.177439 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc26eed77_aa87_4b6b_9f81_aefe3e5ed786.slice/crio-85a3c84f201280c5632dd5f05531f555877d5d4d9b82ca3c2393310e41a88f66 WatchSource:0}: Error finding container 85a3c84f201280c5632dd5f05531f555877d5d4d9b82ca3c2393310e41a88f66: Status 404 returned error can't find the container with id 85a3c84f201280c5632dd5f05531f555877d5d4d9b82ca3c2393310e41a88f66 Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.180155 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.190218 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.202143 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb"] Dec 02 10:35:19 crc kubenswrapper[4679]: W1202 10:35:19.214169 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebbb7257_823e_49a1_af4c_703b979de700.slice/crio-f5459b34c180658751c0678f54899850991e6aba3360b7c59cd6bd6a4f91633b WatchSource:0}: Error finding container f5459b34c180658751c0678f54899850991e6aba3360b7c59cd6bd6a4f91633b: Status 404 returned error can't find the container with id f5459b34c180658751c0678f54899850991e6aba3360b7c59cd6bd6a4f91633b Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.215771 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62"] Dec 02 10:35:19 crc kubenswrapper[4679]: W1202 10:35:19.228892 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ac36ac3_f2ba_40be_9544_9897a27ddaf4.slice/crio-a2aa5fe946586c6f3247dbc81d720354053ff0da6e1231b6ddd9eade55fb9b89 WatchSource:0}: Error finding container a2aa5fe946586c6f3247dbc81d720354053ff0da6e1231b6ddd9eade55fb9b89: Status 404 returned error can't find the container with id a2aa5fe946586c6f3247dbc81d720354053ff0da6e1231b6ddd9eade55fb9b89 Dec 02 10:35:19 crc kubenswrapper[4679]: W1202 10:35:19.231260 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf48fce46_9ae8_4154_b67b_3edd56267209.slice/crio-52ce2cbe57d98b31505974fef0735892086aa017c11e970a0e959276042f1ba9 WatchSource:0}: Error finding container 52ce2cbe57d98b31505974fef0735892086aa017c11e970a0e959276042f1ba9: Status 404 returned error can't find the container with id 52ce2cbe57d98b31505974fef0735892086aa017c11e970a0e959276042f1ba9 Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.231950 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8pdk5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-sxgbw_openstack-operators(ebbb7257-823e-49a1-af4c-703b979de700): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.232496 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bbzgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-lxjdn_openstack-operators(0a119e38-647a-42ab-823f-e5ecea4afbf0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.232962 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gxn8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-ps9nb_openstack-operators(921e63bc-e66b-48d5-a22c-010c65a0e1b0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.234195 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" podUID="0a119e38-647a-42ab-823f-e5ecea4afbf0" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.234781 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gxn8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-ps9nb_openstack-operators(921e63bc-e66b-48d5-a22c-010c65a0e1b0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.234794 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8pdk5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-sxgbw_openstack-operators(ebbb7257-823e-49a1-af4c-703b979de700): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.235049 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j5xr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-7kl62_openstack-operators(f48fce46-9ae8-4154-b67b-3edd56267209): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.236280 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" podUID="ebbb7257-823e-49a1-af4c-703b979de700" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.236352 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" podUID="921e63bc-e66b-48d5-a22c-010c65a0e1b0" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.237696 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j5xr5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-7kl62_openstack-operators(f48fce46-9ae8-4154-b67b-3edd56267209): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.239240 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" podUID="f48fce46-9ae8-4154-b67b-3edd56267209" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.239649 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdtxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-b74hj_openstack-operators(6ac36ac3-f2ba-40be-9544-9897a27ddaf4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.242894 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdtxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-b74hj_openstack-operators(6ac36ac3-f2ba-40be-9544-9897a27ddaf4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.243996 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" podUID="6ac36ac3-f2ba-40be-9544-9897a27ddaf4" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.579099 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.579200 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.579365 4679 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.579411 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:21.579397576 +0000 UTC m=+974.909536436 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "metrics-server-cert" not found Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.579699 4679 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 10:35:19 crc kubenswrapper[4679]: E1202 10:35:19.579725 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:21.579717536 +0000 UTC m=+974.909856396 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "webhook-server-cert" not found Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.846176 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r5dvf"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.848139 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.883456 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-utilities\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.883597 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-catalog-content\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.883716 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghcnj\" (UniqueName: \"kubernetes.io/projected/36f7dfd6-7643-4c2f-895b-0312aeb90a73-kube-api-access-ghcnj\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.904160 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r5dvf"] Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.992945 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghcnj\" (UniqueName: \"kubernetes.io/projected/36f7dfd6-7643-4c2f-895b-0312aeb90a73-kube-api-access-ghcnj\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.993027 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-utilities\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.993086 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-catalog-content\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.994280 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-utilities\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:19 crc kubenswrapper[4679]: I1202 10:35:19.994412 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-catalog-content\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.017453 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghcnj\" (UniqueName: \"kubernetes.io/projected/36f7dfd6-7643-4c2f-895b-0312aeb90a73-kube-api-access-ghcnj\") pod \"community-operators-r5dvf\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.075763 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" event={"ID":"e334fbef-5aab-4cd6-83a1-084160e32d90","Type":"ContainerStarted","Data":"bb5558522586fb0ed057914b65f33153e81e64f7fda9c4422b8bf9f8de33b47b"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.077073 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" event={"ID":"c26eed77-aa87-4b6b-9f81-aefe3e5ed786","Type":"ContainerStarted","Data":"85a3c84f201280c5632dd5f05531f555877d5d4d9b82ca3c2393310e41a88f66"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.079184 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" event={"ID":"ebbb7257-823e-49a1-af4c-703b979de700","Type":"ContainerStarted","Data":"f5459b34c180658751c0678f54899850991e6aba3360b7c59cd6bd6a4f91633b"} Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.084880 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" podUID="ebbb7257-823e-49a1-af4c-703b979de700" Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.089187 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" event={"ID":"9dc9ab59-e368-452c-9ce2-e4d028fd35a1","Type":"ContainerStarted","Data":"949fce7957c4a0ec03d9745da48e081edf269d74f7f81b2be72d8ed41c303732"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.091833 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" event={"ID":"6ac36ac3-f2ba-40be-9544-9897a27ddaf4","Type":"ContainerStarted","Data":"a2aa5fe946586c6f3247dbc81d720354053ff0da6e1231b6ddd9eade55fb9b89"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.095736 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" event={"ID":"1cf6ea83-714b-4a9d-8a5e-708293a3e9e6","Type":"ContainerStarted","Data":"e5ac62a1ab05f229b404e15ff6a08fbff948a4391201bad50641e9d7ab9120bc"} Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.098161 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" podUID="6ac36ac3-f2ba-40be-9544-9897a27ddaf4" Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.111184 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" event={"ID":"0a119e38-647a-42ab-823f-e5ecea4afbf0","Type":"ContainerStarted","Data":"7cfaaefc0f939e3beddd6cc44cdd30a24eba4e7ddd296373fc742c46b8d9dcb9"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.115128 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" event={"ID":"921e63bc-e66b-48d5-a22c-010c65a0e1b0","Type":"ContainerStarted","Data":"b4030bde255defec8eaf83a1ed47a3cbaf3947f62e8e8d0a5a7a327e380506e5"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.118576 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" event={"ID":"f48fce46-9ae8-4154-b67b-3edd56267209","Type":"ContainerStarted","Data":"52ce2cbe57d98b31505974fef0735892086aa017c11e970a0e959276042f1ba9"} Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.120318 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" podUID="0a119e38-647a-42ab-823f-e5ecea4afbf0" Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.121722 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" podUID="f48fce46-9ae8-4154-b67b-3edd56267209" Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.173702 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" event={"ID":"3fa279c2-e6ef-4131-ab48-28306a24000e","Type":"ContainerStarted","Data":"1332acc9d3dd95955259e69bfdc1ae27eed15a27639d13743ad5c96c3110bdb5"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.180100 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.181042 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" podUID="921e63bc-e66b-48d5-a22c-010c65a0e1b0" Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.184623 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" event={"ID":"1b3b0e6e-eb2f-40fc-a77f-c8889816dee3","Type":"ContainerStarted","Data":"7695add452d4f7f667891d3dce951b734f03d212ec973a4cb57b51e0d724069e"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.197666 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" event={"ID":"a877b082-6f7d-4f85-b290-356fe7c8bb5f","Type":"ContainerStarted","Data":"10a80cfc7d5e4608e2438699669a1cd55f4d4aac47cb940ae473279a5fe7ad9a"} Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.713227 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.713578 4679 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:20 crc kubenswrapper[4679]: E1202 10:35:20.713645 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert podName:a0088242-8397-41e7-96b8-d971f307eefc nodeName:}" failed. No retries permitted until 2025-12-02 10:35:24.713623012 +0000 UTC m=+978.043761872 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert") pod "infra-operator-controller-manager-57548d458d-ghgsb" (UID: "a0088242-8397-41e7-96b8-d971f307eefc") : secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:20 crc kubenswrapper[4679]: I1202 10:35:20.979569 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r5dvf"] Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.194527 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" podUID="0a119e38-647a-42ab-823f-e5ecea4afbf0" Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.194788 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" podUID="921e63bc-e66b-48d5-a22c-010c65a0e1b0" Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.196731 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" podUID="ebbb7257-823e-49a1-af4c-703b979de700" Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.200059 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" podUID="f48fce46-9ae8-4154-b67b-3edd56267209" Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.210945 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" podUID="6ac36ac3-f2ba-40be-9544-9897a27ddaf4" Dec 02 10:35:21 crc kubenswrapper[4679]: I1202 10:35:21.234947 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.235606 4679 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.237564 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert podName:b255bbc9-0452-4445-9470-8bd7ad85257a nodeName:}" failed. No retries permitted until 2025-12-02 10:35:25.237525443 +0000 UTC m=+978.567664433 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" (UID: "b255bbc9-0452-4445-9470-8bd7ad85257a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:21 crc kubenswrapper[4679]: I1202 10:35:21.641170 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:21 crc kubenswrapper[4679]: I1202 10:35:21.641285 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.641343 4679 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.641418 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:25.641399235 +0000 UTC m=+978.971538095 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "webhook-server-cert" not found Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.641449 4679 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 10:35:21 crc kubenswrapper[4679]: E1202 10:35:21.641550 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:25.641531099 +0000 UTC m=+978.971669959 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "metrics-server-cert" not found Dec 02 10:35:22 crc kubenswrapper[4679]: W1202 10:35:22.081373 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36f7dfd6_7643_4c2f_895b_0312aeb90a73.slice/crio-a090f364cc17b3bda1aa94919715f35e94e5af312e546f7d9cfab93d32e49702 WatchSource:0}: Error finding container a090f364cc17b3bda1aa94919715f35e94e5af312e546f7d9cfab93d32e49702: Status 404 returned error can't find the container with id a090f364cc17b3bda1aa94919715f35e94e5af312e546f7d9cfab93d32e49702 Dec 02 10:35:22 crc kubenswrapper[4679]: I1202 10:35:22.206584 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5dvf" event={"ID":"36f7dfd6-7643-4c2f-895b-0312aeb90a73","Type":"ContainerStarted","Data":"a090f364cc17b3bda1aa94919715f35e94e5af312e546f7d9cfab93d32e49702"} Dec 02 10:35:24 crc kubenswrapper[4679]: I1202 10:35:24.810059 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:24 crc kubenswrapper[4679]: E1202 10:35:24.810245 4679 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:24 crc kubenswrapper[4679]: E1202 10:35:24.810381 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert podName:a0088242-8397-41e7-96b8-d971f307eefc nodeName:}" failed. No retries permitted until 2025-12-02 10:35:32.810355439 +0000 UTC m=+986.140494369 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert") pod "infra-operator-controller-manager-57548d458d-ghgsb" (UID: "a0088242-8397-41e7-96b8-d971f307eefc") : secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:25 crc kubenswrapper[4679]: I1202 10:35:25.318098 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:25 crc kubenswrapper[4679]: E1202 10:35:25.318444 4679 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:25 crc kubenswrapper[4679]: E1202 10:35:25.318689 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert podName:b255bbc9-0452-4445-9470-8bd7ad85257a nodeName:}" failed. No retries permitted until 2025-12-02 10:35:33.318659371 +0000 UTC m=+986.648798271 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" (UID: "b255bbc9-0452-4445-9470-8bd7ad85257a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 10:35:25 crc kubenswrapper[4679]: I1202 10:35:25.723940 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:25 crc kubenswrapper[4679]: I1202 10:35:25.724059 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:25 crc kubenswrapper[4679]: E1202 10:35:25.724111 4679 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 10:35:25 crc kubenswrapper[4679]: E1202 10:35:25.724205 4679 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 10:35:25 crc kubenswrapper[4679]: E1202 10:35:25.724214 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:33.72419024 +0000 UTC m=+987.054329110 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "metrics-server-cert" not found Dec 02 10:35:25 crc kubenswrapper[4679]: E1202 10:35:25.724262 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:33.724245411 +0000 UTC m=+987.054384291 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "webhook-server-cert" not found Dec 02 10:35:32 crc kubenswrapper[4679]: I1202 10:35:32.832877 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:32 crc kubenswrapper[4679]: E1202 10:35:32.833119 4679 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:32 crc kubenswrapper[4679]: E1202 10:35:32.834003 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert podName:a0088242-8397-41e7-96b8-d971f307eefc nodeName:}" failed. No retries permitted until 2025-12-02 10:35:48.833961603 +0000 UTC m=+1002.164100503 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert") pod "infra-operator-controller-manager-57548d458d-ghgsb" (UID: "a0088242-8397-41e7-96b8-d971f307eefc") : secret "infra-operator-webhook-server-cert" not found Dec 02 10:35:33 crc kubenswrapper[4679]: I1202 10:35:33.339967 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:33 crc kubenswrapper[4679]: I1202 10:35:33.354638 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b255bbc9-0452-4445-9470-8bd7ad85257a-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh\" (UID: \"b255bbc9-0452-4445-9470-8bd7ad85257a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:33 crc kubenswrapper[4679]: I1202 10:35:33.370804 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:33 crc kubenswrapper[4679]: I1202 10:35:33.745238 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:33 crc kubenswrapper[4679]: I1202 10:35:33.745354 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:33 crc kubenswrapper[4679]: E1202 10:35:33.745552 4679 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 10:35:33 crc kubenswrapper[4679]: E1202 10:35:33.745676 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs podName:6edbf2ab-201d-4a7f-902a-7b74566b5a36 nodeName:}" failed. No retries permitted until 2025-12-02 10:35:49.745645202 +0000 UTC m=+1003.075784102 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs") pod "openstack-operator-controller-manager-7bc5fbbb46-rqv9x" (UID: "6edbf2ab-201d-4a7f-902a-7b74566b5a36") : secret "metrics-server-cert" not found Dec 02 10:35:33 crc kubenswrapper[4679]: I1202 10:35:33.752988 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-webhook-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:35 crc kubenswrapper[4679]: E1202 10:35:35.169251 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 02 10:35:35 crc kubenswrapper[4679]: E1202 10:35:35.169807 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5lvdc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-nw7fm_openstack-operators(a877b082-6f7d-4f85-b290-356fe7c8bb5f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:35:35 crc kubenswrapper[4679]: E1202 10:35:35.729913 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 02 10:35:35 crc kubenswrapper[4679]: E1202 10:35:35.730106 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vccbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-nm552_openstack-operators(1b3b0e6e-eb2f-40fc-a77f-c8889816dee3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:35:36 crc kubenswrapper[4679]: E1202 10:35:36.258329 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 02 10:35:36 crc kubenswrapper[4679]: E1202 10:35:36.258489 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4msg8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-jxm5w_openstack-operators(f9cf4107-9e23-4246-a8b7-b3bbda0cc461): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:35:38 crc kubenswrapper[4679]: E1202 10:35:38.195715 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 02 10:35:38 crc kubenswrapper[4679]: E1202 10:35:38.196590 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mvs52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-5zzmd_openstack-operators(5ee03695-d4bb-4b77-84d1-84190ae307d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:35:38 crc kubenswrapper[4679]: E1202 10:35:38.729759 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 02 10:35:38 crc kubenswrapper[4679]: E1202 10:35:38.730138 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fzpw4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-tp7zw_openstack-operators(1cf6ea83-714b-4a9d-8a5e-708293a3e9e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:35:39 crc kubenswrapper[4679]: E1202 10:35:39.386125 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 02 10:35:39 crc kubenswrapper[4679]: E1202 10:35:39.386291 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-th2gd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-628rp_openstack-operators(9dc9ab59-e368-452c-9ce2-e4d028fd35a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:35:43 crc kubenswrapper[4679]: I1202 10:35:43.360942 4679 generic.go:334] "Generic (PLEG): container finished" podID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerID="c3a6252a850d849867c0adfcae00b25607c8b8cc1a84201016c49268cb0e5742" exitCode=0 Dec 02 10:35:43 crc kubenswrapper[4679]: I1202 10:35:43.361007 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5dvf" event={"ID":"36f7dfd6-7643-4c2f-895b-0312aeb90a73","Type":"ContainerDied","Data":"c3a6252a850d849867c0adfcae00b25607c8b8cc1a84201016c49268cb0e5742"} Dec 02 10:35:43 crc kubenswrapper[4679]: I1202 10:35:43.979637 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh"] Dec 02 10:35:44 crc kubenswrapper[4679]: W1202 10:35:44.325760 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb255bbc9_0452_4445_9470_8bd7ad85257a.slice/crio-cc121d0065c4a8da95b5ca4ba8fce44b1db4db28cf478fd14923b24d02ca80b8 WatchSource:0}: Error finding container cc121d0065c4a8da95b5ca4ba8fce44b1db4db28cf478fd14923b24d02ca80b8: Status 404 returned error can't find the container with id cc121d0065c4a8da95b5ca4ba8fce44b1db4db28cf478fd14923b24d02ca80b8 Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.385875 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" event={"ID":"921e63bc-e66b-48d5-a22c-010c65a0e1b0","Type":"ContainerStarted","Data":"ee6b25b8fc998753682d384c9d463348a05ff8c98312d4e459e526acd91b7456"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.406382 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" event={"ID":"e23de616-947b-4ff1-b05b-522a1c39cc7b","Type":"ContainerStarted","Data":"24d8ba7d87f4e900d10ecf2c2238e0dc386f293f197433bc60d61b1b00777403"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.408258 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" event={"ID":"b255bbc9-0452-4445-9470-8bd7ad85257a","Type":"ContainerStarted","Data":"cc121d0065c4a8da95b5ca4ba8fce44b1db4db28cf478fd14923b24d02ca80b8"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.438284 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" event={"ID":"7938c8f2-53f6-4143-997d-a44ea54c9fc2","Type":"ContainerStarted","Data":"aa5d09ad2167813f461c0212ec69cc7e7c3b0e9a46fb10d4f6d305fbbb1dbf5e"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.449684 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" event={"ID":"48f2b621-5d9b-4213-8681-8107d1fc04a4","Type":"ContainerStarted","Data":"c87ffda1a214bf5ffe9b3e3f2ee50e192afec9640083b7aa40dacde0549ff9b6"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.453734 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" event={"ID":"85ed4b5c-eda1-481a-84ae-b97d01a9b9eb","Type":"ContainerStarted","Data":"3c6629f259bcf639bf330deb0987d55c30f8615bd704ee87ff6a390c4b729d63"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.465112 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" event={"ID":"e334fbef-5aab-4cd6-83a1-084160e32d90","Type":"ContainerStarted","Data":"7a9996315a6f1fe53507a3c5af7b032c09f69c85555db7a2e1ed977603157008"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.475326 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" event={"ID":"80f86afb-0d79-4d1c-a0fd-b1d573d3c471","Type":"ContainerStarted","Data":"2e6b0854f851db8840ecad44391ceb84ee26b6e744d96722664d510e537e357c"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.477193 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" event={"ID":"c26eed77-aa87-4b6b-9f81-aefe3e5ed786","Type":"ContainerStarted","Data":"2f3767770918107ab5ba2a2f5273c3b0513fc7abca4bdeb9e8b52a76fdc5049a"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.483263 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" event={"ID":"3fa279c2-e6ef-4131-ab48-28306a24000e","Type":"ContainerStarted","Data":"b92fc7183d552642480412264b45f495d99401e19f9f266e826ac8a81fe228de"} Dec 02 10:35:44 crc kubenswrapper[4679]: I1202 10:35:44.506242 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" event={"ID":"61bf694c-dbb6-442c-aaa0-3c13ff7efc97","Type":"ContainerStarted","Data":"8c80113916aa8d047e027dcbe356139215e9725f282390b40f130e908b41ca5b"} Dec 02 10:35:45 crc kubenswrapper[4679]: I1202 10:35:45.516031 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" event={"ID":"f48fce46-9ae8-4154-b67b-3edd56267209","Type":"ContainerStarted","Data":"156d1ec981700aa9111ed7b3921150b9b5885f9030826f516054ba90859c83e2"} Dec 02 10:35:45 crc kubenswrapper[4679]: I1202 10:35:45.518463 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" event={"ID":"ebbb7257-823e-49a1-af4c-703b979de700","Type":"ContainerStarted","Data":"018e9522664b98c12167c0115487a92004fd7cd286d437217975f39e958f955b"} Dec 02 10:35:46 crc kubenswrapper[4679]: I1202 10:35:46.528555 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" event={"ID":"0a119e38-647a-42ab-823f-e5ecea4afbf0","Type":"ContainerStarted","Data":"4e8403f971e7f4cea19cfce7ab3a4203d59bf1df460062a1b697f98b6af3a9b3"} Dec 02 10:35:46 crc kubenswrapper[4679]: I1202 10:35:46.533426 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" event={"ID":"6ac36ac3-f2ba-40be-9544-9897a27ddaf4","Type":"ContainerStarted","Data":"ad64192b78c9f562bd4ef57a67b4da0f86bd6b2762d01c3214697b87d683e04e"} Dec 02 10:35:46 crc kubenswrapper[4679]: I1202 10:35:46.551940 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lxjdn" podStartSLOduration=5.056682296 podStartE2EDuration="29.551922744s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.232410831 +0000 UTC m=+972.562549701" lastFinishedPulling="2025-12-02 10:35:43.727651289 +0000 UTC m=+997.057790149" observedRunningTime="2025-12-02 10:35:46.546404945 +0000 UTC m=+999.876543825" watchObservedRunningTime="2025-12-02 10:35:46.551922744 +0000 UTC m=+999.882061604" Dec 02 10:35:46 crc kubenswrapper[4679]: I1202 10:35:46.932112 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:35:46 crc kubenswrapper[4679]: I1202 10:35:46.932188 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.575852 4679 generic.go:334] "Generic (PLEG): container finished" podID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerID="5cd5af586371b21c9260312180827b51e19483d212a8d89584c582bc1a12996e" exitCode=0 Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.575983 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5dvf" event={"ID":"36f7dfd6-7643-4c2f-895b-0312aeb90a73","Type":"ContainerDied","Data":"5cd5af586371b21c9260312180827b51e19483d212a8d89584c582bc1a12996e"} Dec 02 10:35:48 crc kubenswrapper[4679]: E1202 10:35:48.580718 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" podUID="1b3b0e6e-eb2f-40fc-a77f-c8889816dee3" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.588209 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" event={"ID":"c26eed77-aa87-4b6b-9f81-aefe3e5ed786","Type":"ContainerStarted","Data":"4ec1c715ceec7c26ce2d9278a010b4d32d0ae4a669e9aef76441369082a39597"} Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.589198 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.591763 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" event={"ID":"b255bbc9-0452-4445-9470-8bd7ad85257a","Type":"ContainerStarted","Data":"f13b84a7a6bcd94292f6eed3991d6892b5470e5a41fabfd08cb8853ef19145c9"} Dec 02 10:35:48 crc kubenswrapper[4679]: E1202 10:35:48.592599 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" podUID="f9cf4107-9e23-4246-a8b7-b3bbda0cc461" Dec 02 10:35:48 crc kubenswrapper[4679]: E1202 10:35:48.602655 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" podUID="9dc9ab59-e368-452c-9ce2-e4d028fd35a1" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.636536 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" podStartSLOduration=2.801936123 podStartE2EDuration="31.63628363s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.19850271 +0000 UTC m=+972.528641570" lastFinishedPulling="2025-12-02 10:35:48.032850217 +0000 UTC m=+1001.362989077" observedRunningTime="2025-12-02 10:35:48.63350025 +0000 UTC m=+1001.963639110" watchObservedRunningTime="2025-12-02 10:35:48.63628363 +0000 UTC m=+1001.966422480" Dec 02 10:35:48 crc kubenswrapper[4679]: E1202 10:35:48.659250 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" podUID="1cf6ea83-714b-4a9d-8a5e-708293a3e9e6" Dec 02 10:35:48 crc kubenswrapper[4679]: E1202 10:35:48.835552 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" podUID="5ee03695-d4bb-4b77-84d1-84190ae307d6" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.921599 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.929181 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0088242-8397-41e7-96b8-d971f307eefc-cert\") pod \"infra-operator-controller-manager-57548d458d-ghgsb\" (UID: \"a0088242-8397-41e7-96b8-d971f307eefc\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:48 crc kubenswrapper[4679]: I1202 10:35:48.955741 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:48 crc kubenswrapper[4679]: E1202 10:35:48.964398 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" podUID="a877b082-6f7d-4f85-b290-356fe7c8bb5f" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.523808 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb"] Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.607109 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" event={"ID":"1cf6ea83-714b-4a9d-8a5e-708293a3e9e6","Type":"ContainerStarted","Data":"38d2330da954f4606e102e9b9fb49a3d9ec4acbb2de75042907ee4b0ebf7d1f5"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.611705 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" event={"ID":"7938c8f2-53f6-4143-997d-a44ea54c9fc2","Type":"ContainerStarted","Data":"7fb03635960448089c037e0ea3895ebbd0b92766de2560b9eac71b39e6dcc2bd"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.612582 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.618086 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.620024 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" event={"ID":"48f2b621-5d9b-4213-8681-8107d1fc04a4","Type":"ContainerStarted","Data":"2dfb84df4b114ae15b69d309d2ee4bd2e74de03a9cc5470ebcc438118843afd2"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.620653 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.622156 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" event={"ID":"6ac36ac3-f2ba-40be-9544-9897a27ddaf4","Type":"ContainerStarted","Data":"9d551d903bceb974fc32b1981dd731324ff85707631f60d14dafb38594b2c9d7"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.622508 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.627013 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.630999 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" event={"ID":"e23de616-947b-4ff1-b05b-522a1c39cc7b","Type":"ContainerStarted","Data":"87d859364e4af139ad0147266e3f3a9428dc1ae7d83abc4b661e7b096352a2e5"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.631865 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.639518 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.646458 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" event={"ID":"ebbb7257-823e-49a1-af4c-703b979de700","Type":"ContainerStarted","Data":"fddba7609f0250ad65594cd0c9fa156d249c5fb4d5f3f4a3312b5233d8d6eff0"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.646946 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.649477 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.649867 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" event={"ID":"1b3b0e6e-eb2f-40fc-a77f-c8889816dee3","Type":"ContainerStarted","Data":"3f44bef05148e43a9b12735a469066957b5b751c2063f6c2d4d2283960aece74"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.658585 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" event={"ID":"e334fbef-5aab-4cd6-83a1-084160e32d90","Type":"ContainerStarted","Data":"d6f22534b513619326bd3d5bd9fa126d2cd0b07094bf2de92e92e2970c654a91"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.659397 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.664774 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-48bsh" podStartSLOduration=4.37923021 podStartE2EDuration="33.664762747s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.749277517 +0000 UTC m=+972.079416377" lastFinishedPulling="2025-12-02 10:35:48.034810054 +0000 UTC m=+1001.364948914" observedRunningTime="2025-12-02 10:35:49.650508865 +0000 UTC m=+1002.980647725" watchObservedRunningTime="2025-12-02 10:35:49.664762747 +0000 UTC m=+1002.994901607" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.669401 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.669929 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" event={"ID":"80f86afb-0d79-4d1c-a0fd-b1d573d3c471","Type":"ContainerStarted","Data":"0306c2d73a558c4607a220f76b37bc35c6ec3b3cecf807423e03c652773d1d0f"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.670857 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.680023 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-snss6" podStartSLOduration=3.42011711 podStartE2EDuration="33.680012919s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:17.834010524 +0000 UTC m=+971.164149384" lastFinishedPulling="2025-12-02 10:35:48.093906333 +0000 UTC m=+1001.424045193" observedRunningTime="2025-12-02 10:35:49.676279261 +0000 UTC m=+1003.006418151" watchObservedRunningTime="2025-12-02 10:35:49.680012919 +0000 UTC m=+1003.010151779" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.680485 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.680631 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" event={"ID":"f9cf4107-9e23-4246-a8b7-b3bbda0cc461","Type":"ContainerStarted","Data":"4e87d03d204a2abab2307a50db0fe03c280430f090c0a26b2646f4f8481144e1"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.688908 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" event={"ID":"3fa279c2-e6ef-4131-ab48-28306a24000e","Type":"ContainerStarted","Data":"953d79d9d697644c99d8f4670369564b6baf469308315604b1ad7d6742f04366"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.689718 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.693908 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" event={"ID":"9dc9ab59-e368-452c-9ce2-e4d028fd35a1","Type":"ContainerStarted","Data":"39fc65a4d8f356eaf6d71353f196ee93b469096bc54cf862955f2a7114a30012"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.695173 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" event={"ID":"a877b082-6f7d-4f85-b290-356fe7c8bb5f","Type":"ContainerStarted","Data":"9781e7e11a275fb15c8628351d73cc605500f6a179313906d8d2d5f48fa57200"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.697137 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.715632 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" event={"ID":"85ed4b5c-eda1-481a-84ae-b97d01a9b9eb","Type":"ContainerStarted","Data":"2b81521c4ed88a1087d6b03b104c370f6a371f33c227a80ac113255051178cd0"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.715918 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.720370 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" podStartSLOduration=3.864889316 podStartE2EDuration="32.720355925s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.239527596 +0000 UTC m=+972.569666456" lastFinishedPulling="2025-12-02 10:35:48.094994205 +0000 UTC m=+1001.425133065" observedRunningTime="2025-12-02 10:35:49.719723627 +0000 UTC m=+1003.049862487" watchObservedRunningTime="2025-12-02 10:35:49.720355925 +0000 UTC m=+1003.050494785" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.723610 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.730626 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" event={"ID":"921e63bc-e66b-48d5-a22c-010c65a0e1b0","Type":"ContainerStarted","Data":"df904b748259ee1250b8df30b10619dbf0bff69170aa843273b5825573d438fa"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.746726 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.759337 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6edbf2ab-201d-4a7f-902a-7b74566b5a36-metrics-certs\") pod \"openstack-operator-controller-manager-7bc5fbbb46-rqv9x\" (UID: \"6edbf2ab-201d-4a7f-902a-7b74566b5a36\") " pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.767578 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.770994 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jmj2d" podStartSLOduration=3.925026023 podStartE2EDuration="33.770975129s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.188683153 +0000 UTC m=+971.518822013" lastFinishedPulling="2025-12-02 10:35:48.034632259 +0000 UTC m=+1001.364771119" observedRunningTime="2025-12-02 10:35:49.759360553 +0000 UTC m=+1003.089499433" watchObservedRunningTime="2025-12-02 10:35:49.770975129 +0000 UTC m=+1003.101113989" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.785679 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" event={"ID":"b255bbc9-0452-4445-9470-8bd7ad85257a","Type":"ContainerStarted","Data":"ac09e3ede11f2099b98edf3c6971ae011fea7b0d715ba1f481f3346e35aaa601"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.786384 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.811964 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" event={"ID":"f48fce46-9ae8-4154-b67b-3edd56267209","Type":"ContainerStarted","Data":"e85effcb9ea5926675a89297b21d3a8770995fe52e0777177d907acc90a294f7"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.813928 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.821196 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.827775 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" event={"ID":"5ee03695-d4bb-4b77-84d1-84190ae307d6","Type":"ContainerStarted","Data":"7fa120b7deaa02dce65216fe28ff6b169d3966ccd2e649e3bfee16b0035c8121"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.843175 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" podStartSLOduration=3.836246567 podStartE2EDuration="32.843156786s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.232778851 +0000 UTC m=+972.562917711" lastFinishedPulling="2025-12-02 10:35:48.23968907 +0000 UTC m=+1001.569827930" observedRunningTime="2025-12-02 10:35:49.839687936 +0000 UTC m=+1003.169826796" watchObservedRunningTime="2025-12-02 10:35:49.843156786 +0000 UTC m=+1003.173295636" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.879881 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" event={"ID":"61bf694c-dbb6-442c-aaa0-3c13ff7efc97","Type":"ContainerStarted","Data":"a26bce4f4ade78039e002b4266e10495cd6abfeb3cf5441026342f51d4da5ca5"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.880929 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.889903 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.895713 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" event={"ID":"a0088242-8397-41e7-96b8-d971f307eefc","Type":"ContainerStarted","Data":"0e52e138ca2c88cb300551fbacbc5ed7a8a9e2aaa4a31bddfa6375a9f2318d8b"} Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.912530 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-9flx2" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.927862 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" podStartSLOduration=4.386106549 podStartE2EDuration="33.927847056s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.231829034 +0000 UTC m=+972.561967894" lastFinishedPulling="2025-12-02 10:35:48.773569541 +0000 UTC m=+1002.103708401" observedRunningTime="2025-12-02 10:35:49.926247389 +0000 UTC m=+1003.256386249" watchObservedRunningTime="2025-12-02 10:35:49.927847056 +0000 UTC m=+1003.257985916" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.950121 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bnz2w" podStartSLOduration=3.91771707 podStartE2EDuration="33.950102839s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.741821401 +0000 UTC m=+972.071960261" lastFinishedPulling="2025-12-02 10:35:48.77420717 +0000 UTC m=+1002.104346030" observedRunningTime="2025-12-02 10:35:49.948823582 +0000 UTC m=+1003.278962442" watchObservedRunningTime="2025-12-02 10:35:49.950102839 +0000 UTC m=+1003.280241699" Dec 02 10:35:49 crc kubenswrapper[4679]: I1202 10:35:49.986395 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-w74h9" podStartSLOduration=4.406105046 podStartE2EDuration="33.986377508s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.151709586 +0000 UTC m=+972.481848446" lastFinishedPulling="2025-12-02 10:35:48.731982048 +0000 UTC m=+1002.062120908" observedRunningTime="2025-12-02 10:35:49.983606838 +0000 UTC m=+1003.313745708" watchObservedRunningTime="2025-12-02 10:35:49.986377508 +0000 UTC m=+1003.316516368" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.077599 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xfz4c" podStartSLOduration=3.819691199 podStartE2EDuration="33.077560626s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.163987642 +0000 UTC m=+972.494126502" lastFinishedPulling="2025-12-02 10:35:48.421857069 +0000 UTC m=+1001.751995929" observedRunningTime="2025-12-02 10:35:50.055904169 +0000 UTC m=+1003.386043029" watchObservedRunningTime="2025-12-02 10:35:50.077560626 +0000 UTC m=+1003.407699486" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.102702 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-59k6f" podStartSLOduration=4.681225073 podStartE2EDuration="34.102682852s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.616072604 +0000 UTC m=+971.946211464" lastFinishedPulling="2025-12-02 10:35:48.037530383 +0000 UTC m=+1001.367669243" observedRunningTime="2025-12-02 10:35:50.100289873 +0000 UTC m=+1003.430428743" watchObservedRunningTime="2025-12-02 10:35:50.102682852 +0000 UTC m=+1003.432821712" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.246684 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" podStartSLOduration=29.616469941 podStartE2EDuration="33.246663057s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:44.339609828 +0000 UTC m=+997.669748688" lastFinishedPulling="2025-12-02 10:35:47.969802944 +0000 UTC m=+1001.299941804" observedRunningTime="2025-12-02 10:35:50.241794196 +0000 UTC m=+1003.571933066" watchObservedRunningTime="2025-12-02 10:35:50.246663057 +0000 UTC m=+1003.576801907" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.292206 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-7kl62" podStartSLOduration=4.116947156 podStartE2EDuration="33.292186383s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.234958274 +0000 UTC m=+972.565097144" lastFinishedPulling="2025-12-02 10:35:48.410197511 +0000 UTC m=+1001.740336371" observedRunningTime="2025-12-02 10:35:50.285272563 +0000 UTC m=+1003.615411423" watchObservedRunningTime="2025-12-02 10:35:50.292186383 +0000 UTC m=+1003.622325243" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.324458 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-442vr" podStartSLOduration=4.942337796 podStartE2EDuration="34.324441246s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.652124657 +0000 UTC m=+971.982263517" lastFinishedPulling="2025-12-02 10:35:48.034228107 +0000 UTC m=+1001.364366967" observedRunningTime="2025-12-02 10:35:50.319796522 +0000 UTC m=+1003.649935382" watchObservedRunningTime="2025-12-02 10:35:50.324441246 +0000 UTC m=+1003.654580106" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.361619 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x"] Dec 02 10:35:50 crc kubenswrapper[4679]: W1202 10:35:50.409948 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6edbf2ab_201d_4a7f_902a_7b74566b5a36.slice/crio-3092aea4d1c897b11617dfa388b46cd7d9ec30475c6bd02e7c07b2943383b00e WatchSource:0}: Error finding container 3092aea4d1c897b11617dfa388b46cd7d9ec30475c6bd02e7c07b2943383b00e: Status 404 returned error can't find the container with id 3092aea4d1c897b11617dfa388b46cd7d9ec30475c6bd02e7c07b2943383b00e Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.947846 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" event={"ID":"5ee03695-d4bb-4b77-84d1-84190ae307d6","Type":"ContainerStarted","Data":"c5b5c8a05eef845347d26da9022fa302e5b67633feca3a56f02bc889b9897e77"} Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.947921 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.960327 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5dvf" event={"ID":"36f7dfd6-7643-4c2f-895b-0312aeb90a73","Type":"ContainerStarted","Data":"22a83dd1bf585d1a3f6ec3d319814b9107908f923323083df4eb75d98c5d78ec"} Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.965460 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" podStartSLOduration=2.972130052 podStartE2EDuration="34.965438626s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.730930466 +0000 UTC m=+972.061069326" lastFinishedPulling="2025-12-02 10:35:50.72423904 +0000 UTC m=+1004.054377900" observedRunningTime="2025-12-02 10:35:50.960205494 +0000 UTC m=+1004.290344354" watchObservedRunningTime="2025-12-02 10:35:50.965438626 +0000 UTC m=+1004.295577606" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.980592 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" event={"ID":"1b3b0e6e-eb2f-40fc-a77f-c8889816dee3","Type":"ContainerStarted","Data":"df75313d6abf240937305938133a0b396baef3a5ef42f86d329d3301d827b4aa"} Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.981488 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.983107 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r5dvf" podStartSLOduration=26.211789704 podStartE2EDuration="31.983098847s" podCreationTimestamp="2025-12-02 10:35:19 +0000 UTC" firstStartedPulling="2025-12-02 10:35:43.523264347 +0000 UTC m=+996.853403247" lastFinishedPulling="2025-12-02 10:35:49.29457353 +0000 UTC m=+1002.624712390" observedRunningTime="2025-12-02 10:35:50.981679306 +0000 UTC m=+1004.311818166" watchObservedRunningTime="2025-12-02 10:35:50.983098847 +0000 UTC m=+1004.313237697" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.991774 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" event={"ID":"9dc9ab59-e368-452c-9ce2-e4d028fd35a1","Type":"ContainerStarted","Data":"c9809c9548e8a5c612b99fd80bdf2bdc70a452dda450320134c734fc06c20c15"} Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.991818 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.994840 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" event={"ID":"a877b082-6f7d-4f85-b290-356fe7c8bb5f","Type":"ContainerStarted","Data":"1fc86ea373439ab3657f442ab9e6b208f2a72f82cff423a9d75799a96a02ea43"} Dec 02 10:35:50 crc kubenswrapper[4679]: I1202 10:35:50.994877 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.003682 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" event={"ID":"1cf6ea83-714b-4a9d-8a5e-708293a3e9e6","Type":"ContainerStarted","Data":"6e7e4459fbab8b31323e03cb8f69b42489602c490f9dbb296c5c11dbe9403f94"} Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.003778 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.009523 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" event={"ID":"6edbf2ab-201d-4a7f-902a-7b74566b5a36","Type":"ContainerStarted","Data":"e1e2be0cfb5a689da81a510ad0ab7ad23baaa3885a3497b6ac5eca05efd62971"} Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.009554 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" event={"ID":"6edbf2ab-201d-4a7f-902a-7b74566b5a36","Type":"ContainerStarted","Data":"3092aea4d1c897b11617dfa388b46cd7d9ec30475c6bd02e7c07b2943383b00e"} Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.009551 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" podStartSLOduration=2.95292326 podStartE2EDuration="34.009531901s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.217394546 +0000 UTC m=+972.547533406" lastFinishedPulling="2025-12-02 10:35:50.274003187 +0000 UTC m=+1003.604142047" observedRunningTime="2025-12-02 10:35:51.002597801 +0000 UTC m=+1004.332736661" watchObservedRunningTime="2025-12-02 10:35:51.009531901 +0000 UTC m=+1004.339670761" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.010077 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.018784 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" event={"ID":"f9cf4107-9e23-4246-a8b7-b3bbda0cc461","Type":"ContainerStarted","Data":"3ee46dfe7cf95b26be1a6fcf7ace5dabee773bd7fd6492de1ac5de278b6f5869"} Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.018987 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.019842 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" podStartSLOduration=2.787553147 podStartE2EDuration="34.019828139s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.177468451 +0000 UTC m=+972.507607311" lastFinishedPulling="2025-12-02 10:35:50.409743433 +0000 UTC m=+1003.739882303" observedRunningTime="2025-12-02 10:35:51.017557243 +0000 UTC m=+1004.347696103" watchObservedRunningTime="2025-12-02 10:35:51.019828139 +0000 UTC m=+1004.349966999" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.025515 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-b74hj" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.033737 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-ps9nb" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.053108 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" podStartSLOduration=2.497975851 podStartE2EDuration="34.053085791s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.170480409 +0000 UTC m=+972.500619269" lastFinishedPulling="2025-12-02 10:35:50.725590349 +0000 UTC m=+1004.055729209" observedRunningTime="2025-12-02 10:35:51.048253581 +0000 UTC m=+1004.378392441" watchObservedRunningTime="2025-12-02 10:35:51.053085791 +0000 UTC m=+1004.383224651" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.110190 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" podStartSLOduration=3.230121004 podStartE2EDuration="35.110172702s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:18.746521467 +0000 UTC m=+972.076660327" lastFinishedPulling="2025-12-02 10:35:50.626573165 +0000 UTC m=+1003.956712025" observedRunningTime="2025-12-02 10:35:51.109900064 +0000 UTC m=+1004.440038924" watchObservedRunningTime="2025-12-02 10:35:51.110172702 +0000 UTC m=+1004.440311562" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.136954 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" podStartSLOduration=4.075017571 podStartE2EDuration="35.136932256s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:19.15116074 +0000 UTC m=+972.481299600" lastFinishedPulling="2025-12-02 10:35:50.213075425 +0000 UTC m=+1003.543214285" observedRunningTime="2025-12-02 10:35:51.134805584 +0000 UTC m=+1004.464944444" watchObservedRunningTime="2025-12-02 10:35:51.136932256 +0000 UTC m=+1004.467071126" Dec 02 10:35:51 crc kubenswrapper[4679]: I1202 10:35:51.215092 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" podStartSLOduration=34.215072516 podStartE2EDuration="34.215072516s" podCreationTimestamp="2025-12-02 10:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:35:51.206024374 +0000 UTC m=+1004.536163234" watchObservedRunningTime="2025-12-02 10:35:51.215072516 +0000 UTC m=+1004.545211376" Dec 02 10:35:52 crc kubenswrapper[4679]: I1202 10:35:52.028978 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:53 crc kubenswrapper[4679]: I1202 10:35:53.035174 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" event={"ID":"a0088242-8397-41e7-96b8-d971f307eefc","Type":"ContainerStarted","Data":"6cd91b568ea4b9a1358a252daf7c42809b64291d1dcf578bc8190124bad7332c"} Dec 02 10:35:53 crc kubenswrapper[4679]: I1202 10:35:53.035563 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" event={"ID":"a0088242-8397-41e7-96b8-d971f307eefc","Type":"ContainerStarted","Data":"18bc1d91b531f9d1e77f20fff73d4106c2446e1a4335338cd564338b1a58c2e7"} Dec 02 10:35:53 crc kubenswrapper[4679]: I1202 10:35:53.056519 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" podStartSLOduration=33.910766183 podStartE2EDuration="37.056499746s" podCreationTimestamp="2025-12-02 10:35:16 +0000 UTC" firstStartedPulling="2025-12-02 10:35:49.537169267 +0000 UTC m=+1002.867308127" lastFinishedPulling="2025-12-02 10:35:52.68290283 +0000 UTC m=+1006.013041690" observedRunningTime="2025-12-02 10:35:53.053845669 +0000 UTC m=+1006.383984539" watchObservedRunningTime="2025-12-02 10:35:53.056499746 +0000 UTC m=+1006.386638596" Dec 02 10:35:53 crc kubenswrapper[4679]: I1202 10:35:53.381802 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh" Dec 02 10:35:54 crc kubenswrapper[4679]: I1202 10:35:54.055978 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:57 crc kubenswrapper[4679]: I1202 10:35:57.361030 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jxm5w" Dec 02 10:35:57 crc kubenswrapper[4679]: I1202 10:35:57.452963 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-5zzmd" Dec 02 10:35:57 crc kubenswrapper[4679]: I1202 10:35:57.500934 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-tp7zw" Dec 02 10:35:57 crc kubenswrapper[4679]: I1202 10:35:57.599376 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-628rp" Dec 02 10:35:57 crc kubenswrapper[4679]: I1202 10:35:57.717705 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-nm552" Dec 02 10:35:57 crc kubenswrapper[4679]: I1202 10:35:57.930637 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nw7fm" Dec 02 10:35:58 crc kubenswrapper[4679]: I1202 10:35:58.961851 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-ghgsb" Dec 02 10:35:59 crc kubenswrapper[4679]: I1202 10:35:59.777543 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7bc5fbbb46-rqv9x" Dec 02 10:36:00 crc kubenswrapper[4679]: I1202 10:36:00.180590 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:36:00 crc kubenswrapper[4679]: I1202 10:36:00.180647 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:36:00 crc kubenswrapper[4679]: I1202 10:36:00.231035 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:36:01 crc kubenswrapper[4679]: I1202 10:36:01.158414 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:36:01 crc kubenswrapper[4679]: I1202 10:36:01.202559 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r5dvf"] Dec 02 10:36:03 crc kubenswrapper[4679]: I1202 10:36:03.131906 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r5dvf" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="registry-server" containerID="cri-o://22a83dd1bf585d1a3f6ec3d319814b9107908f923323083df4eb75d98c5d78ec" gracePeriod=2 Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.271781 4679 generic.go:334] "Generic (PLEG): container finished" podID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerID="22a83dd1bf585d1a3f6ec3d319814b9107908f923323083df4eb75d98c5d78ec" exitCode=0 Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.271878 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5dvf" event={"ID":"36f7dfd6-7643-4c2f-895b-0312aeb90a73","Type":"ContainerDied","Data":"22a83dd1bf585d1a3f6ec3d319814b9107908f923323083df4eb75d98c5d78ec"} Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.726832 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.850093 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-catalog-content\") pod \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.850174 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghcnj\" (UniqueName: \"kubernetes.io/projected/36f7dfd6-7643-4c2f-895b-0312aeb90a73-kube-api-access-ghcnj\") pod \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.850218 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-utilities\") pod \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\" (UID: \"36f7dfd6-7643-4c2f-895b-0312aeb90a73\") " Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.851135 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-utilities" (OuterVolumeSpecName: "utilities") pod "36f7dfd6-7643-4c2f-895b-0312aeb90a73" (UID: "36f7dfd6-7643-4c2f-895b-0312aeb90a73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.861633 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36f7dfd6-7643-4c2f-895b-0312aeb90a73-kube-api-access-ghcnj" (OuterVolumeSpecName: "kube-api-access-ghcnj") pod "36f7dfd6-7643-4c2f-895b-0312aeb90a73" (UID: "36f7dfd6-7643-4c2f-895b-0312aeb90a73"). InnerVolumeSpecName "kube-api-access-ghcnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.952151 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.952293 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghcnj\" (UniqueName: \"kubernetes.io/projected/36f7dfd6-7643-4c2f-895b-0312aeb90a73-kube-api-access-ghcnj\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.968472 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36f7dfd6-7643-4c2f-895b-0312aeb90a73" (UID: "36f7dfd6-7643-4c2f-895b-0312aeb90a73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.970547 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-87j6s"] Dec 02 10:36:08 crc kubenswrapper[4679]: E1202 10:36:08.970931 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="extract-utilities" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.970944 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="extract-utilities" Dec 02 10:36:08 crc kubenswrapper[4679]: E1202 10:36:08.970958 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="registry-server" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.970964 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="registry-server" Dec 02 10:36:08 crc kubenswrapper[4679]: E1202 10:36:08.970990 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="extract-content" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.970997 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="extract-content" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.971297 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" containerName="registry-server" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.972577 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:08 crc kubenswrapper[4679]: I1202 10:36:08.988763 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-87j6s"] Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.053942 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w57r\" (UniqueName: \"kubernetes.io/projected/96633b7d-7894-4e3e-adf2-ce16371793f5-kube-api-access-6w57r\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.054008 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-catalog-content\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.054061 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-utilities\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.054183 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36f7dfd6-7643-4c2f-895b-0312aeb90a73-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.155277 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w57r\" (UniqueName: \"kubernetes.io/projected/96633b7d-7894-4e3e-adf2-ce16371793f5-kube-api-access-6w57r\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.155362 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-catalog-content\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.155388 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-utilities\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.155872 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-utilities\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.155961 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-catalog-content\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.173186 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w57r\" (UniqueName: \"kubernetes.io/projected/96633b7d-7894-4e3e-adf2-ce16371793f5-kube-api-access-6w57r\") pod \"certified-operators-87j6s\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.280914 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5dvf" event={"ID":"36f7dfd6-7643-4c2f-895b-0312aeb90a73","Type":"ContainerDied","Data":"a090f364cc17b3bda1aa94919715f35e94e5af312e546f7d9cfab93d32e49702"} Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.280972 4679 scope.go:117] "RemoveContainer" containerID="22a83dd1bf585d1a3f6ec3d319814b9107908f923323083df4eb75d98c5d78ec" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.280987 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5dvf" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.293641 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.305018 4679 scope.go:117] "RemoveContainer" containerID="5cd5af586371b21c9260312180827b51e19483d212a8d89584c582bc1a12996e" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.322389 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r5dvf"] Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.333367 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r5dvf"] Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.350533 4679 scope.go:117] "RemoveContainer" containerID="c3a6252a850d849867c0adfcae00b25607c8b8cc1a84201016c49268cb0e5742" Dec 02 10:36:09 crc kubenswrapper[4679]: I1202 10:36:09.806295 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-87j6s"] Dec 02 10:36:09 crc kubenswrapper[4679]: W1202 10:36:09.810345 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96633b7d_7894_4e3e_adf2_ce16371793f5.slice/crio-33fd0e88661cce2e8a89ba9d4a58c0099c3030bc2449c5effe1ff7febbcf1fc7 WatchSource:0}: Error finding container 33fd0e88661cce2e8a89ba9d4a58c0099c3030bc2449c5effe1ff7febbcf1fc7: Status 404 returned error can't find the container with id 33fd0e88661cce2e8a89ba9d4a58c0099c3030bc2449c5effe1ff7febbcf1fc7 Dec 02 10:36:10 crc kubenswrapper[4679]: I1202 10:36:10.289916 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87j6s" event={"ID":"96633b7d-7894-4e3e-adf2-ce16371793f5","Type":"ContainerStarted","Data":"33fd0e88661cce2e8a89ba9d4a58c0099c3030bc2449c5effe1ff7febbcf1fc7"} Dec 02 10:36:10 crc kubenswrapper[4679]: I1202 10:36:10.919769 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36f7dfd6-7643-4c2f-895b-0312aeb90a73" path="/var/lib/kubelet/pods/36f7dfd6-7643-4c2f-895b-0312aeb90a73/volumes" Dec 02 10:36:12 crc kubenswrapper[4679]: I1202 10:36:12.307754 4679 generic.go:334] "Generic (PLEG): container finished" podID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerID="aedc29723859c6db3659c11b1742c0b85a913a26863325ce342cfd547585c041" exitCode=0 Dec 02 10:36:12 crc kubenswrapper[4679]: I1202 10:36:12.307935 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87j6s" event={"ID":"96633b7d-7894-4e3e-adf2-ce16371793f5","Type":"ContainerDied","Data":"aedc29723859c6db3659c11b1742c0b85a913a26863325ce342cfd547585c041"} Dec 02 10:36:12 crc kubenswrapper[4679]: I1202 10:36:12.310188 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:36:14 crc kubenswrapper[4679]: I1202 10:36:14.324577 4679 generic.go:334] "Generic (PLEG): container finished" podID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerID="3c84167258d351233d5a9663d5976bd1480bc1088c74bea75ef18ef9feaff791" exitCode=0 Dec 02 10:36:14 crc kubenswrapper[4679]: I1202 10:36:14.324703 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87j6s" event={"ID":"96633b7d-7894-4e3e-adf2-ce16371793f5","Type":"ContainerDied","Data":"3c84167258d351233d5a9663d5976bd1480bc1088c74bea75ef18ef9feaff791"} Dec 02 10:36:15 crc kubenswrapper[4679]: I1202 10:36:15.333770 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87j6s" event={"ID":"96633b7d-7894-4e3e-adf2-ce16371793f5","Type":"ContainerStarted","Data":"bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb"} Dec 02 10:36:15 crc kubenswrapper[4679]: I1202 10:36:15.351725 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-87j6s" podStartSLOduration=4.835478876 podStartE2EDuration="7.351709063s" podCreationTimestamp="2025-12-02 10:36:08 +0000 UTC" firstStartedPulling="2025-12-02 10:36:12.309974448 +0000 UTC m=+1025.640113308" lastFinishedPulling="2025-12-02 10:36:14.826204635 +0000 UTC m=+1028.156343495" observedRunningTime="2025-12-02 10:36:15.349330164 +0000 UTC m=+1028.679469024" watchObservedRunningTime="2025-12-02 10:36:15.351709063 +0000 UTC m=+1028.681847913" Dec 02 10:36:16 crc kubenswrapper[4679]: I1202 10:36:16.932124 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:36:16 crc kubenswrapper[4679]: I1202 10:36:16.932990 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:36:19 crc kubenswrapper[4679]: I1202 10:36:19.294488 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:19 crc kubenswrapper[4679]: I1202 10:36:19.294888 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:19 crc kubenswrapper[4679]: I1202 10:36:19.347760 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.959132 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jzzth"] Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.961256 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.963524 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.963544 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.963776 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-55r2h" Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.967927 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 02 10:36:20 crc kubenswrapper[4679]: I1202 10:36:20.972968 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jzzth"] Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.026780 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cgzx\" (UniqueName: \"kubernetes.io/projected/191f4986-63bd-4114-b8af-d499f87c83ae-kube-api-access-5cgzx\") pod \"dnsmasq-dns-675f4bcbfc-jzzth\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.026874 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/191f4986-63bd-4114-b8af-d499f87c83ae-config\") pod \"dnsmasq-dns-675f4bcbfc-jzzth\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.031116 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkn7t"] Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.032733 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.040206 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.048991 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkn7t"] Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.128360 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cgzx\" (UniqueName: \"kubernetes.io/projected/191f4986-63bd-4114-b8af-d499f87c83ae-kube-api-access-5cgzx\") pod \"dnsmasq-dns-675f4bcbfc-jzzth\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.128548 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/191f4986-63bd-4114-b8af-d499f87c83ae-config\") pod \"dnsmasq-dns-675f4bcbfc-jzzth\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.129553 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/191f4986-63bd-4114-b8af-d499f87c83ae-config\") pod \"dnsmasq-dns-675f4bcbfc-jzzth\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.158092 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cgzx\" (UniqueName: \"kubernetes.io/projected/191f4986-63bd-4114-b8af-d499f87c83ae-kube-api-access-5cgzx\") pod \"dnsmasq-dns-675f4bcbfc-jzzth\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.229545 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-config\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.229706 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms7gk\" (UniqueName: \"kubernetes.io/projected/cdf10f73-5189-42a8-8aa4-95c5a19fb351-kube-api-access-ms7gk\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.229790 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.305054 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.330922 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms7gk\" (UniqueName: \"kubernetes.io/projected/cdf10f73-5189-42a8-8aa4-95c5a19fb351-kube-api-access-ms7gk\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.331005 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.331091 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-config\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.331984 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-config\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.332889 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.360228 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms7gk\" (UniqueName: \"kubernetes.io/projected/cdf10f73-5189-42a8-8aa4-95c5a19fb351-kube-api-access-ms7gk\") pod \"dnsmasq-dns-78dd6ddcc-xkn7t\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.655097 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.750165 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jzzth"] Dec 02 10:36:21 crc kubenswrapper[4679]: W1202 10:36:21.758564 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod191f4986_63bd_4114_b8af_d499f87c83ae.slice/crio-bcfc2ca11709ccdbb2e46bd1f89ae9e1f5ccf2242b193a189d69f39407c208ba WatchSource:0}: Error finding container bcfc2ca11709ccdbb2e46bd1f89ae9e1f5ccf2242b193a189d69f39407c208ba: Status 404 returned error can't find the container with id bcfc2ca11709ccdbb2e46bd1f89ae9e1f5ccf2242b193a189d69f39407c208ba Dec 02 10:36:21 crc kubenswrapper[4679]: W1202 10:36:21.898662 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdf10f73_5189_42a8_8aa4_95c5a19fb351.slice/crio-6d674408ac63bb33ff5258c3353ac0abe7335926ec2fd58b89086f56f1941b9f WatchSource:0}: Error finding container 6d674408ac63bb33ff5258c3353ac0abe7335926ec2fd58b89086f56f1941b9f: Status 404 returned error can't find the container with id 6d674408ac63bb33ff5258c3353ac0abe7335926ec2fd58b89086f56f1941b9f Dec 02 10:36:21 crc kubenswrapper[4679]: I1202 10:36:21.900265 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkn7t"] Dec 02 10:36:22 crc kubenswrapper[4679]: I1202 10:36:22.374626 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" event={"ID":"191f4986-63bd-4114-b8af-d499f87c83ae","Type":"ContainerStarted","Data":"bcfc2ca11709ccdbb2e46bd1f89ae9e1f5ccf2242b193a189d69f39407c208ba"} Dec 02 10:36:22 crc kubenswrapper[4679]: I1202 10:36:22.377102 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" event={"ID":"cdf10f73-5189-42a8-8aa4-95c5a19fb351","Type":"ContainerStarted","Data":"6d674408ac63bb33ff5258c3353ac0abe7335926ec2fd58b89086f56f1941b9f"} Dec 02 10:36:23 crc kubenswrapper[4679]: I1202 10:36:23.994784 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jzzth"] Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.028401 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sv9tw"] Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.030144 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.079026 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sv9tw"] Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.184378 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.184457 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-config\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.184502 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrx6b\" (UniqueName: \"kubernetes.io/projected/227e410d-8f47-41ac-9f78-be5cbeb6c16f-kube-api-access-jrx6b\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.287188 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrx6b\" (UniqueName: \"kubernetes.io/projected/227e410d-8f47-41ac-9f78-be5cbeb6c16f-kube-api-access-jrx6b\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.287272 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.287330 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-config\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.288170 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-config\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.288389 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.311032 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkn7t"] Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.311593 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrx6b\" (UniqueName: \"kubernetes.io/projected/227e410d-8f47-41ac-9f78-be5cbeb6c16f-kube-api-access-jrx6b\") pod \"dnsmasq-dns-666b6646f7-sv9tw\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.344269 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zj2hk"] Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.346287 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.352512 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zj2hk"] Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.371466 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.494064 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-config\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.494114 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghhjk\" (UniqueName: \"kubernetes.io/projected/f27f0728-a9c2-4ce6-85ea-dab897da5f35-kube-api-access-ghhjk\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.494139 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.597515 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-config\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.597555 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghhjk\" (UniqueName: \"kubernetes.io/projected/f27f0728-a9c2-4ce6-85ea-dab897da5f35-kube-api-access-ghhjk\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.597583 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.600006 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.601519 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-config\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.625281 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghhjk\" (UniqueName: \"kubernetes.io/projected/f27f0728-a9c2-4ce6-85ea-dab897da5f35-kube-api-access-ghhjk\") pod \"dnsmasq-dns-57d769cc4f-zj2hk\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.668183 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:36:24 crc kubenswrapper[4679]: I1202 10:36:24.699860 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sv9tw"] Dec 02 10:36:24 crc kubenswrapper[4679]: W1202 10:36:24.729356 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod227e410d_8f47_41ac_9f78_be5cbeb6c16f.slice/crio-4a6b549897c34aa8a3162b6c05eb43daa63b1eb3c8f876b7145d86b9c9d1e93b WatchSource:0}: Error finding container 4a6b549897c34aa8a3162b6c05eb43daa63b1eb3c8f876b7145d86b9c9d1e93b: Status 404 returned error can't find the container with id 4a6b549897c34aa8a3162b6c05eb43daa63b1eb3c8f876b7145d86b9c9d1e93b Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.175278 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.176559 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.178822 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.179173 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.179682 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.179999 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.180170 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.180488 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2dqtv" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.180628 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.186027 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.215286 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zj2hk"] Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309060 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-pod-info\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309100 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309131 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309407 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309484 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-config-data\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309503 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309528 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309554 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-server-conf\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309658 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309751 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w6pp\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-kube-api-access-8w6pp\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.309779 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.410974 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411013 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-config-data\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411033 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411053 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411073 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-server-conf\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411093 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411117 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w6pp\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-kube-api-access-8w6pp\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411132 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411163 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-pod-info\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411180 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411206 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411552 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.411876 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.412054 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.412819 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.415570 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-server-conf\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.436040 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-pod-info\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.436814 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.437006 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-config-data\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.442371 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w6pp\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-kube-api-access-8w6pp\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.446473 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.454192 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.459579 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" event={"ID":"227e410d-8f47-41ac-9f78-be5cbeb6c16f","Type":"ContainerStarted","Data":"4a6b549897c34aa8a3162b6c05eb43daa63b1eb3c8f876b7145d86b9c9d1e93b"} Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.474988 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" event={"ID":"f27f0728-a9c2-4ce6-85ea-dab897da5f35","Type":"ContainerStarted","Data":"4be41f6f21d92f148243e348473862465ab22d915cea586d6da903cc2cce810c"} Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.490686 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.508646 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.514216 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.515762 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.522574 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.522876 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-78jjf" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.522995 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.523122 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.523233 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.527599 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.529222 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.537114 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.618199 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.619484 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.627202 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.629558 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.629688 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.629844 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6nl2\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-kube-api-access-d6nl2\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.630443 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.630593 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.630779 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.630891 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.630987 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732555 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732624 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732663 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732726 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732777 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732810 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732835 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732865 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732890 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6nl2\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-kube-api-access-d6nl2\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732941 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.732978 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.736487 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.736679 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.737790 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.738886 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.739245 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.741019 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.741455 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.745095 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.751637 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.760857 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.765074 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6nl2\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-kube-api-access-d6nl2\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.772949 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:25 crc kubenswrapper[4679]: I1202 10:36:25.984794 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.102451 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:36:26 crc kubenswrapper[4679]: W1202 10:36:26.135456 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66a56dd2_b866_4b5b_9fd5_94e3dc8422ae.slice/crio-01c828f2474a6ee870a17d1e49a3ca2a84c45a85740075857e43ac83bb0cc680 WatchSource:0}: Error finding container 01c828f2474a6ee870a17d1e49a3ca2a84c45a85740075857e43ac83bb0cc680: Status 404 returned error can't find the container with id 01c828f2474a6ee870a17d1e49a3ca2a84c45a85740075857e43ac83bb0cc680 Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.483025 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae","Type":"ContainerStarted","Data":"01c828f2474a6ee870a17d1e49a3ca2a84c45a85740075857e43ac83bb0cc680"} Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.578129 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:36:26 crc kubenswrapper[4679]: W1202 10:36:26.582526 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod416d04f1_7afb_4bdf_a5d5_1a0c381513f9.slice/crio-df03b283b858a9fd75c51bacb1b6f3af4a33b66d4389397c7dd1422fed752005 WatchSource:0}: Error finding container df03b283b858a9fd75c51bacb1b6f3af4a33b66d4389397c7dd1422fed752005: Status 404 returned error can't find the container with id df03b283b858a9fd75c51bacb1b6f3af4a33b66d4389397c7dd1422fed752005 Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.899234 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.905131 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.909184 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.915260 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.916573 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.917597 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-b6s2c" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.924944 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.954404 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1f514bf-70bf-484c-b485-7e3d46a5c139-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.954928 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.954983 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1f514bf-70bf-484c-b485-7e3d46a5c139-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.955030 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.955076 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxhfx\" (UniqueName: \"kubernetes.io/projected/e1f514bf-70bf-484c-b485-7e3d46a5c139-kube-api-access-wxhfx\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.955203 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f514bf-70bf-484c-b485-7e3d46a5c139-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.955267 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-config-data-default\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.955296 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-kolla-config\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:26 crc kubenswrapper[4679]: I1202 10:36:26.958514 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.058574 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.058623 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1f514bf-70bf-484c-b485-7e3d46a5c139-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.058657 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.058695 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxhfx\" (UniqueName: \"kubernetes.io/projected/e1f514bf-70bf-484c-b485-7e3d46a5c139-kube-api-access-wxhfx\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.059047 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f514bf-70bf-484c-b485-7e3d46a5c139-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.059120 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-config-data-default\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.059187 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-kolla-config\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.059263 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1f514bf-70bf-484c-b485-7e3d46a5c139-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.060333 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.060941 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-config-data-default\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.059152 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e1f514bf-70bf-484c-b485-7e3d46a5c139-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.060383 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.061217 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e1f514bf-70bf-484c-b485-7e3d46a5c139-kolla-config\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.078828 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1f514bf-70bf-484c-b485-7e3d46a5c139-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.100325 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1f514bf-70bf-484c-b485-7e3d46a5c139-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.111570 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.111731 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxhfx\" (UniqueName: \"kubernetes.io/projected/e1f514bf-70bf-484c-b485-7e3d46a5c139-kube-api-access-wxhfx\") pod \"openstack-galera-0\" (UID: \"e1f514bf-70bf-484c-b485-7e3d46a5c139\") " pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.255339 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.520045 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"416d04f1-7afb-4bdf-a5d5-1a0c381513f9","Type":"ContainerStarted","Data":"df03b283b858a9fd75c51bacb1b6f3af4a33b66d4389397c7dd1422fed752005"} Dec 02 10:36:27 crc kubenswrapper[4679]: I1202 10:36:27.780022 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.185711 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.203691 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.203730 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.206888 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-l9r75" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.206966 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.207234 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.208201 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.297295 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.297411 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqk5\" (UniqueName: \"kubernetes.io/projected/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-kube-api-access-zsqk5\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.297485 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.297802 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.297916 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.297993 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.298266 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.298411 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.399802 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.399863 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.399910 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.399937 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.399991 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.400031 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqk5\" (UniqueName: \"kubernetes.io/projected/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-kube-api-access-zsqk5\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.400079 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.400136 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.400714 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.401390 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.401488 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.402033 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.402146 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.424119 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.424210 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.435738 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqk5\" (UniqueName: \"kubernetes.io/projected/30ab65a3-79d4-4eb7-a6b3-c759422ddffc-kube-api-access-zsqk5\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.448683 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"30ab65a3-79d4-4eb7-a6b3-c759422ddffc\") " pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.534651 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.540660 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e1f514bf-70bf-484c-b485-7e3d46a5c139","Type":"ContainerStarted","Data":"8c5e378c280cb9165e3cbfc81f9418060edc0d485f9bda00e0d0cff7eff3d2f1"} Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.569661 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.570618 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.575017 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.575089 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-h9vf2" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.575674 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.593269 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.609535 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltbf6\" (UniqueName: \"kubernetes.io/projected/afdf0aae-7bec-4c47-820f-5a970c264069-kube-api-access-ltbf6\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.609614 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afdf0aae-7bec-4c47-820f-5a970c264069-config-data\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.609683 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdf0aae-7bec-4c47-820f-5a970c264069-combined-ca-bundle\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.609718 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdf0aae-7bec-4c47-820f-5a970c264069-memcached-tls-certs\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.609745 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/afdf0aae-7bec-4c47-820f-5a970c264069-kolla-config\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.712077 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdf0aae-7bec-4c47-820f-5a970c264069-combined-ca-bundle\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.712132 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdf0aae-7bec-4c47-820f-5a970c264069-memcached-tls-certs\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.712157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/afdf0aae-7bec-4c47-820f-5a970c264069-kolla-config\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.712197 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltbf6\" (UniqueName: \"kubernetes.io/projected/afdf0aae-7bec-4c47-820f-5a970c264069-kube-api-access-ltbf6\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.712233 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afdf0aae-7bec-4c47-820f-5a970c264069-config-data\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.716770 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdf0aae-7bec-4c47-820f-5a970c264069-memcached-tls-certs\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.721074 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/afdf0aae-7bec-4c47-820f-5a970c264069-kolla-config\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.721103 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afdf0aae-7bec-4c47-820f-5a970c264069-config-data\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.723845 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdf0aae-7bec-4c47-820f-5a970c264069-combined-ca-bundle\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.750870 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltbf6\" (UniqueName: \"kubernetes.io/projected/afdf0aae-7bec-4c47-820f-5a970c264069-kube-api-access-ltbf6\") pod \"memcached-0\" (UID: \"afdf0aae-7bec-4c47-820f-5a970c264069\") " pod="openstack/memcached-0" Dec 02 10:36:28 crc kubenswrapper[4679]: I1202 10:36:28.912226 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 02 10:36:29 crc kubenswrapper[4679]: I1202 10:36:29.209823 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 10:36:29 crc kubenswrapper[4679]: I1202 10:36:29.381356 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:29 crc kubenswrapper[4679]: I1202 10:36:29.429377 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-87j6s"] Dec 02 10:36:29 crc kubenswrapper[4679]: I1202 10:36:29.479646 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 02 10:36:29 crc kubenswrapper[4679]: I1202 10:36:29.548521 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-87j6s" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="registry-server" containerID="cri-o://bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb" gracePeriod=2 Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.570082 4679 generic.go:334] "Generic (PLEG): container finished" podID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerID="bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb" exitCode=0 Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.570574 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87j6s" event={"ID":"96633b7d-7894-4e3e-adf2-ce16371793f5","Type":"ContainerDied","Data":"bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb"} Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.605030 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.607145 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.608761 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-ttx85" Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.613886 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.649130 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqtbj\" (UniqueName: \"kubernetes.io/projected/e2a75d52-ee01-460e-b15b-70bff6593ef1-kube-api-access-lqtbj\") pod \"kube-state-metrics-0\" (UID: \"e2a75d52-ee01-460e-b15b-70bff6593ef1\") " pod="openstack/kube-state-metrics-0" Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.750140 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqtbj\" (UniqueName: \"kubernetes.io/projected/e2a75d52-ee01-460e-b15b-70bff6593ef1-kube-api-access-lqtbj\") pod \"kube-state-metrics-0\" (UID: \"e2a75d52-ee01-460e-b15b-70bff6593ef1\") " pod="openstack/kube-state-metrics-0" Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.770879 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqtbj\" (UniqueName: \"kubernetes.io/projected/e2a75d52-ee01-460e-b15b-70bff6593ef1-kube-api-access-lqtbj\") pod \"kube-state-metrics-0\" (UID: \"e2a75d52-ee01-460e-b15b-70bff6593ef1\") " pod="openstack/kube-state-metrics-0" Dec 02 10:36:30 crc kubenswrapper[4679]: I1202 10:36:30.941496 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.627640 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p5dsw"] Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.628920 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.631074 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bxvdw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.632082 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.635392 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-z6rx5"] Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.637016 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.638164 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.646991 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p5dsw"] Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.654946 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-z6rx5"] Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693162 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-log-ovn\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693208 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55txt\" (UniqueName: \"kubernetes.io/projected/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-kube-api-access-55txt\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693252 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-lib\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693273 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-scripts\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693287 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cbqb\" (UniqueName: \"kubernetes.io/projected/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-kube-api-access-7cbqb\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693330 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-scripts\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693361 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-ovn-controller-tls-certs\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693378 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-run\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693401 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-run\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693424 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-combined-ca-bundle\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693441 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-log\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693459 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-etc-ovs\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.693635 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-run-ovn\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794435 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-lib\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794486 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-scripts\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794513 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cbqb\" (UniqueName: \"kubernetes.io/projected/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-kube-api-access-7cbqb\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794544 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-scripts\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794579 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-ovn-controller-tls-certs\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794601 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-run\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794629 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-run\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794661 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-combined-ca-bundle\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794693 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-log\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794729 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-etc-ovs\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794762 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-run-ovn\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794800 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-log-ovn\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.794852 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55txt\" (UniqueName: \"kubernetes.io/projected/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-kube-api-access-55txt\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795036 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-lib\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795170 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-run\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795371 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-var-log\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795575 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-etc-ovs\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795621 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-run\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795729 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-log-ovn\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.795855 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-var-run-ovn\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.796982 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-scripts\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.797758 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-scripts\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.803424 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-combined-ca-bundle\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.805063 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-ovn-controller-tls-certs\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.812006 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55txt\" (UniqueName: \"kubernetes.io/projected/9a9d4b4d-2f9c-4f27-9f64-98efad54fded-kube-api-access-55txt\") pod \"ovn-controller-ovs-z6rx5\" (UID: \"9a9d4b4d-2f9c-4f27-9f64-98efad54fded\") " pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.813391 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cbqb\" (UniqueName: \"kubernetes.io/projected/9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb-kube-api-access-7cbqb\") pod \"ovn-controller-p5dsw\" (UID: \"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb\") " pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.951203 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:33 crc kubenswrapper[4679]: I1202 10:36:33.978781 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.511226 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.512610 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.533880 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.534010 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.536834 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.539327 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-j8lbt" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.539684 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.539888 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.709544 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-config\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710131 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710280 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710388 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710470 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710551 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710635 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m7zm\" (UniqueName: \"kubernetes.io/projected/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-kube-api-access-8m7zm\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.710731 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812112 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812176 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m7zm\" (UniqueName: \"kubernetes.io/projected/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-kube-api-access-8m7zm\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812219 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812252 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-config\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812321 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812368 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812391 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812412 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.812607 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.813490 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.813640 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.814466 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-config\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.818496 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.819753 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.823738 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.834278 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m7zm\" (UniqueName: \"kubernetes.io/projected/71d9df94-9e39-409b-aa9e-e13a8f8f9ea2-kube-api-access-8m7zm\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.838540 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2\") " pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:34 crc kubenswrapper[4679]: I1202 10:36:34.844656 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 02 10:36:36 crc kubenswrapper[4679]: W1202 10:36:36.549978 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30ab65a3_79d4_4eb7_a6b3_c759422ddffc.slice/crio-0d5d1597466d2b40e7622780f0e66ca73b683b331b6fdee5ffff3b84012f66ba WatchSource:0}: Error finding container 0d5d1597466d2b40e7622780f0e66ca73b683b331b6fdee5ffff3b84012f66ba: Status 404 returned error can't find the container with id 0d5d1597466d2b40e7622780f0e66ca73b683b331b6fdee5ffff3b84012f66ba Dec 02 10:36:36 crc kubenswrapper[4679]: I1202 10:36:36.628382 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"30ab65a3-79d4-4eb7-a6b3-c759422ddffc","Type":"ContainerStarted","Data":"0d5d1597466d2b40e7622780f0e66ca73b683b331b6fdee5ffff3b84012f66ba"} Dec 02 10:36:37 crc kubenswrapper[4679]: W1202 10:36:37.207609 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdf0aae_7bec_4c47_820f_5a970c264069.slice/crio-cd12a307e030bdc831531399eaa30b9c93881eae7455b58a2fd47220ad31bb36 WatchSource:0}: Error finding container cd12a307e030bdc831531399eaa30b9c93881eae7455b58a2fd47220ad31bb36: Status 404 returned error can't find the container with id cd12a307e030bdc831531399eaa30b9c93881eae7455b58a2fd47220ad31bb36 Dec 02 10:36:37 crc kubenswrapper[4679]: I1202 10:36:37.635494 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"afdf0aae-7bec-4c47-820f-5a970c264069","Type":"ContainerStarted","Data":"cd12a307e030bdc831531399eaa30b9c93881eae7455b58a2fd47220ad31bb36"} Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.272333 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.275371 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.288067 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.288375 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.289869 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.290297 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tj9cb" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.296037 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.465442 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.465601 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.471952 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhfmv\" (UniqueName: \"kubernetes.io/projected/285ae6d4-044a-4720-a495-8ca9eca431f3-kube-api-access-zhfmv\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.472976 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/285ae6d4-044a-4720-a495-8ca9eca431f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.473058 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.473083 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.473169 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/285ae6d4-044a-4720-a495-8ca9eca431f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.473397 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/285ae6d4-044a-4720-a495-8ca9eca431f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575021 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575089 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575114 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhfmv\" (UniqueName: \"kubernetes.io/projected/285ae6d4-044a-4720-a495-8ca9eca431f3-kube-api-access-zhfmv\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575137 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/285ae6d4-044a-4720-a495-8ca9eca431f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575156 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575181 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575217 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/285ae6d4-044a-4720-a495-8ca9eca431f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.575273 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/285ae6d4-044a-4720-a495-8ca9eca431f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.576023 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.576077 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/285ae6d4-044a-4720-a495-8ca9eca431f3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.576296 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/285ae6d4-044a-4720-a495-8ca9eca431f3-config\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.577241 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/285ae6d4-044a-4720-a495-8ca9eca431f3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.585000 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.585802 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.594686 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285ae6d4-044a-4720-a495-8ca9eca431f3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.595097 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhfmv\" (UniqueName: \"kubernetes.io/projected/285ae6d4-044a-4720-a495-8ca9eca431f3-kube-api-access-zhfmv\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.598677 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"285ae6d4-044a-4720-a495-8ca9eca431f3\") " pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:38 crc kubenswrapper[4679]: I1202 10:36:38.610558 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 02 10:36:39 crc kubenswrapper[4679]: E1202 10:36:39.294854 4679 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb is running failed: container process not found" containerID="bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb" cmd=["grpc_health_probe","-addr=:50051"] Dec 02 10:36:39 crc kubenswrapper[4679]: E1202 10:36:39.295797 4679 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb is running failed: container process not found" containerID="bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb" cmd=["grpc_health_probe","-addr=:50051"] Dec 02 10:36:39 crc kubenswrapper[4679]: E1202 10:36:39.296356 4679 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb is running failed: container process not found" containerID="bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb" cmd=["grpc_health_probe","-addr=:50051"] Dec 02 10:36:39 crc kubenswrapper[4679]: E1202 10:36:39.296414 4679 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-87j6s" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="registry-server" Dec 02 10:36:46 crc kubenswrapper[4679]: I1202 10:36:46.931495 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:36:46 crc kubenswrapper[4679]: I1202 10:36:46.932278 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:36:46 crc kubenswrapper[4679]: I1202 10:36:46.935469 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:36:46 crc kubenswrapper[4679]: I1202 10:36:46.936777 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0ace29b98cd54337ce3c65db2564b1430008dec833a4152e92f6248a82bf4d3"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:36:46 crc kubenswrapper[4679]: I1202 10:36:46.936853 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://d0ace29b98cd54337ce3c65db2564b1430008dec833a4152e92f6248a82bf4d3" gracePeriod=600 Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.612517 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.719127 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87j6s" event={"ID":"96633b7d-7894-4e3e-adf2-ce16371793f5","Type":"ContainerDied","Data":"33fd0e88661cce2e8a89ba9d4a58c0099c3030bc2449c5effe1ff7febbcf1fc7"} Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.719176 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87j6s" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.719188 4679 scope.go:117] "RemoveContainer" containerID="bd3008a9530e6c668c458dfd6c802f06fc2c6648b6e9d405bbdfac3198064adb" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.722056 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="d0ace29b98cd54337ce3c65db2564b1430008dec833a4152e92f6248a82bf4d3" exitCode=0 Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.722100 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"d0ace29b98cd54337ce3c65db2564b1430008dec833a4152e92f6248a82bf4d3"} Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.750335 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w57r\" (UniqueName: \"kubernetes.io/projected/96633b7d-7894-4e3e-adf2-ce16371793f5-kube-api-access-6w57r\") pod \"96633b7d-7894-4e3e-adf2-ce16371793f5\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.750507 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-catalog-content\") pod \"96633b7d-7894-4e3e-adf2-ce16371793f5\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.750534 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-utilities\") pod \"96633b7d-7894-4e3e-adf2-ce16371793f5\" (UID: \"96633b7d-7894-4e3e-adf2-ce16371793f5\") " Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.751428 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-utilities" (OuterVolumeSpecName: "utilities") pod "96633b7d-7894-4e3e-adf2-ce16371793f5" (UID: "96633b7d-7894-4e3e-adf2-ce16371793f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.756270 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96633b7d-7894-4e3e-adf2-ce16371793f5-kube-api-access-6w57r" (OuterVolumeSpecName: "kube-api-access-6w57r") pod "96633b7d-7894-4e3e-adf2-ce16371793f5" (UID: "96633b7d-7894-4e3e-adf2-ce16371793f5"). InnerVolumeSpecName "kube-api-access-6w57r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.795455 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96633b7d-7894-4e3e-adf2-ce16371793f5" (UID: "96633b7d-7894-4e3e-adf2-ce16371793f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.852050 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w57r\" (UniqueName: \"kubernetes.io/projected/96633b7d-7894-4e3e-adf2-ce16371793f5-kube-api-access-6w57r\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.852086 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:48 crc kubenswrapper[4679]: I1202 10:36:48.852099 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96633b7d-7894-4e3e-adf2-ce16371793f5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:49 crc kubenswrapper[4679]: I1202 10:36:49.050191 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-87j6s"] Dec 02 10:36:49 crc kubenswrapper[4679]: I1202 10:36:49.058347 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-87j6s"] Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.286354 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.286534 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wxhfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(e1f514bf-70bf-484c-b485-7e3d46a5c139): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.287781 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="e1f514bf-70bf-484c-b485-7e3d46a5c139" Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.739114 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="e1f514bf-70bf-484c-b485-7e3d46a5c139" Dec 02 10:36:50 crc kubenswrapper[4679]: I1202 10:36:50.919378 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" path="/var/lib/kubelet/pods/96633b7d-7894-4e3e-adf2-ce16371793f5/volumes" Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.996299 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.996491 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5cgzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-jzzth_openstack(191f4986-63bd-4114-b8af-d499f87c83ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:50 crc kubenswrapper[4679]: E1202 10:36:50.997679 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" podUID="191f4986-63bd-4114-b8af-d499f87c83ae" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.959908 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.960553 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w6pp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(66a56dd2-b866-4b5b-9fd5-94e3dc8422ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.961747 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.977837 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.978009 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d6nl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(416d04f1-7afb-4bdf-a5d5-1a0c381513f9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.979182 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.995623 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.995789 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrx6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-sv9tw_openstack(227e410d-8f47-41ac-9f78-be5cbeb6c16f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:51 crc kubenswrapper[4679]: E1202 10:36:51.997147 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" podUID="227e410d-8f47-41ac-9f78-be5cbeb6c16f" Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.010753 4679 scope.go:117] "RemoveContainer" containerID="3c84167258d351233d5a9663d5976bd1480bc1088c74bea75ef18ef9feaff791" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.012228 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.012366 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ghhjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-zj2hk_openstack(f27f0728-a9c2-4ce6-85ea-dab897da5f35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.014185 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" podUID="f27f0728-a9c2-4ce6-85ea-dab897da5f35" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.034373 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.034550 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ms7gk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-xkn7t_openstack(cdf10f73-5189-42a8-8aa4-95c5a19fb351): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.035951 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" podUID="cdf10f73-5189-42a8-8aa4-95c5a19fb351" Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.756648 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" event={"ID":"191f4986-63bd-4114-b8af-d499f87c83ae","Type":"ContainerDied","Data":"bcfc2ca11709ccdbb2e46bd1f89ae9e1f5ccf2242b193a189d69f39407c208ba"} Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.756837 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcfc2ca11709ccdbb2e46bd1f89ae9e1f5ccf2242b193a189d69f39407c208ba" Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.803638 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.840107 4679 scope.go:117] "RemoveContainer" containerID="aedc29723859c6db3659c11b1742c0b85a913a26863325ce342cfd547585c041" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.840158 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.840404 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" podUID="227e410d-8f47-41ac-9f78-be5cbeb6c16f" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.840471 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" Dec 02 10:36:52 crc kubenswrapper[4679]: E1202 10:36:52.840578 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" podUID="f27f0728-a9c2-4ce6-85ea-dab897da5f35" Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.925099 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cgzx\" (UniqueName: \"kubernetes.io/projected/191f4986-63bd-4114-b8af-d499f87c83ae-kube-api-access-5cgzx\") pod \"191f4986-63bd-4114-b8af-d499f87c83ae\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.925503 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/191f4986-63bd-4114-b8af-d499f87c83ae-config\") pod \"191f4986-63bd-4114-b8af-d499f87c83ae\" (UID: \"191f4986-63bd-4114-b8af-d499f87c83ae\") " Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.926874 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/191f4986-63bd-4114-b8af-d499f87c83ae-config" (OuterVolumeSpecName: "config") pod "191f4986-63bd-4114-b8af-d499f87c83ae" (UID: "191f4986-63bd-4114-b8af-d499f87c83ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:36:52 crc kubenswrapper[4679]: I1202 10:36:52.936149 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/191f4986-63bd-4114-b8af-d499f87c83ae-kube-api-access-5cgzx" (OuterVolumeSpecName: "kube-api-access-5cgzx") pod "191f4986-63bd-4114-b8af-d499f87c83ae" (UID: "191f4986-63bd-4114-b8af-d499f87c83ae"). InnerVolumeSpecName "kube-api-access-5cgzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.028244 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/191f4986-63bd-4114-b8af-d499f87c83ae-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.028658 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cgzx\" (UniqueName: \"kubernetes.io/projected/191f4986-63bd-4114-b8af-d499f87c83ae-kube-api-access-5cgzx\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.033417 4679 scope.go:117] "RemoveContainer" containerID="09c15e7820d4c5f93b890b6f7afc7d06c4c33bed2214883f94f5eae437e987f2" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.255954 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.419235 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.436701 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms7gk\" (UniqueName: \"kubernetes.io/projected/cdf10f73-5189-42a8-8aa4-95c5a19fb351-kube-api-access-ms7gk\") pod \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.436758 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-config\") pod \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.436811 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-dns-svc\") pod \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\" (UID: \"cdf10f73-5189-42a8-8aa4-95c5a19fb351\") " Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.437463 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-config" (OuterVolumeSpecName: "config") pod "cdf10f73-5189-42a8-8aa4-95c5a19fb351" (UID: "cdf10f73-5189-42a8-8aa4-95c5a19fb351"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.438077 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdf10f73-5189-42a8-8aa4-95c5a19fb351" (UID: "cdf10f73-5189-42a8-8aa4-95c5a19fb351"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.438153 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.443065 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf10f73-5189-42a8-8aa4-95c5a19fb351-kube-api-access-ms7gk" (OuterVolumeSpecName: "kube-api-access-ms7gk") pod "cdf10f73-5189-42a8-8aa4-95c5a19fb351" (UID: "cdf10f73-5189-42a8-8aa4-95c5a19fb351"). InnerVolumeSpecName "kube-api-access-ms7gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.473776 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.505955 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p5dsw"] Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.539722 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms7gk\" (UniqueName: \"kubernetes.io/projected/cdf10f73-5189-42a8-8aa4-95c5a19fb351-kube-api-access-ms7gk\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.539769 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf10f73-5189-42a8-8aa4-95c5a19fb351-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.581686 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 10:36:53 crc kubenswrapper[4679]: W1202 10:36:53.582820 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285ae6d4_044a_4720_a495_8ca9eca431f3.slice/crio-b877a0d45a78680c8d59431f6683bc331a6b564b73634af47066163e5c8fe8ca WatchSource:0}: Error finding container b877a0d45a78680c8d59431f6683bc331a6b564b73634af47066163e5c8fe8ca: Status 404 returned error can't find the container with id b877a0d45a78680c8d59431f6683bc331a6b564b73634af47066163e5c8fe8ca Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.763726 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"285ae6d4-044a-4720-a495-8ca9eca431f3","Type":"ContainerStarted","Data":"b877a0d45a78680c8d59431f6683bc331a6b564b73634af47066163e5c8fe8ca"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.765846 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p5dsw" event={"ID":"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb","Type":"ContainerStarted","Data":"0c042152bb297405112cd0c303ff8ae7790c8de60579d28633058d1776b0132e"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.767317 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"30ab65a3-79d4-4eb7-a6b3-c759422ddffc","Type":"ContainerStarted","Data":"930d6af45cffd1439156069204735e22b29fbe4458f315f9048083561c0e80ef"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.769518 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"1c09445f76d33fdda8f22af126b3018e23a8b609e3131d69d4172f788feafb09"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.770450 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2","Type":"ContainerStarted","Data":"9bbafffce37f5033bfbbf2afcadbab4b87e70315f6bf4bb63545673806feb805"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.771247 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2a75d52-ee01-460e-b15b-70bff6593ef1","Type":"ContainerStarted","Data":"89819e526360acedcc60e7831a7b9c8ceccf084b4d96617cd33ee97cc05ebd6c"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.772243 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"afdf0aae-7bec-4c47-820f-5a970c264069","Type":"ContainerStarted","Data":"865a00fc78d753eac30f30bf1053f787d7781bf498479f19bfb9b7e46147dabb"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.772943 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.773874 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jzzth" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.774456 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.774485 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xkn7t" event={"ID":"cdf10f73-5189-42a8-8aa4-95c5a19fb351","Type":"ContainerDied","Data":"6d674408ac63bb33ff5258c3353ac0abe7335926ec2fd58b89086f56f1941b9f"} Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.824000 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.185914395 podStartE2EDuration="25.823977586s" podCreationTimestamp="2025-12-02 10:36:28 +0000 UTC" firstStartedPulling="2025-12-02 10:36:37.209949454 +0000 UTC m=+1050.540088314" lastFinishedPulling="2025-12-02 10:36:52.848012645 +0000 UTC m=+1066.178151505" observedRunningTime="2025-12-02 10:36:53.823336998 +0000 UTC m=+1067.153475868" watchObservedRunningTime="2025-12-02 10:36:53.823977586 +0000 UTC m=+1067.154116446" Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.869794 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkn7t"] Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.876265 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xkn7t"] Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.904214 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jzzth"] Dec 02 10:36:53 crc kubenswrapper[4679]: I1202 10:36:53.912590 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jzzth"] Dec 02 10:36:54 crc kubenswrapper[4679]: I1202 10:36:54.258502 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-z6rx5"] Dec 02 10:36:54 crc kubenswrapper[4679]: I1202 10:36:54.782548 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z6rx5" event={"ID":"9a9d4b4d-2f9c-4f27-9f64-98efad54fded","Type":"ContainerStarted","Data":"d0136866821386a0dd83e56396de0072266626014c8451b6574e7615620da8a6"} Dec 02 10:36:54 crc kubenswrapper[4679]: I1202 10:36:54.919887 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="191f4986-63bd-4114-b8af-d499f87c83ae" path="/var/lib/kubelet/pods/191f4986-63bd-4114-b8af-d499f87c83ae/volumes" Dec 02 10:36:54 crc kubenswrapper[4679]: I1202 10:36:54.920503 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf10f73-5189-42a8-8aa4-95c5a19fb351" path="/var/lib/kubelet/pods/cdf10f73-5189-42a8-8aa4-95c5a19fb351/volumes" Dec 02 10:36:57 crc kubenswrapper[4679]: I1202 10:36:57.807591 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"285ae6d4-044a-4720-a495-8ca9eca431f3","Type":"ContainerStarted","Data":"92fe11bdb62cb0330952363f34e491799d92921bb79cbffbc52b91c1e12ba0d5"} Dec 02 10:36:57 crc kubenswrapper[4679]: I1202 10:36:57.812421 4679 generic.go:334] "Generic (PLEG): container finished" podID="30ab65a3-79d4-4eb7-a6b3-c759422ddffc" containerID="930d6af45cffd1439156069204735e22b29fbe4458f315f9048083561c0e80ef" exitCode=0 Dec 02 10:36:57 crc kubenswrapper[4679]: I1202 10:36:57.812499 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"30ab65a3-79d4-4eb7-a6b3-c759422ddffc","Type":"ContainerDied","Data":"930d6af45cffd1439156069204735e22b29fbe4458f315f9048083561c0e80ef"} Dec 02 10:36:57 crc kubenswrapper[4679]: I1202 10:36:57.817612 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2","Type":"ContainerStarted","Data":"13164f92533597005070588e4d8ab99a44260c8c64d2139d81cca92b81d92e49"} Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.829714 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p5dsw" event={"ID":"9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb","Type":"ContainerStarted","Data":"05ccd2ad31f4a61451bbbdffc608a2f31693a6d23622dbde90dc75c0833cd108"} Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.830129 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-p5dsw" Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.833005 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z6rx5" event={"ID":"9a9d4b4d-2f9c-4f27-9f64-98efad54fded","Type":"ContainerStarted","Data":"70b4b33f544f64dc764e7ec9caa370bfb11b5738c33f8ad8201a22eb12e11b67"} Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.838148 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"30ab65a3-79d4-4eb7-a6b3-c759422ddffc","Type":"ContainerStarted","Data":"2d7049be36755bd7b6691cd771fb450e000db8d7dbc03a79afc08661e299f97f"} Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.853872 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-p5dsw" podStartSLOduration=21.88237621 podStartE2EDuration="25.853852781s" podCreationTimestamp="2025-12-02 10:36:33 +0000 UTC" firstStartedPulling="2025-12-02 10:36:53.520409249 +0000 UTC m=+1066.850548109" lastFinishedPulling="2025-12-02 10:36:57.49188582 +0000 UTC m=+1070.822024680" observedRunningTime="2025-12-02 10:36:58.848987831 +0000 UTC m=+1072.179126691" watchObservedRunningTime="2025-12-02 10:36:58.853852781 +0000 UTC m=+1072.183991641" Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.904169 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.614654277 podStartE2EDuration="31.904150869s" podCreationTimestamp="2025-12-02 10:36:27 +0000 UTC" firstStartedPulling="2025-12-02 10:36:36.554666862 +0000 UTC m=+1049.884805712" lastFinishedPulling="2025-12-02 10:36:52.844163444 +0000 UTC m=+1066.174302304" observedRunningTime="2025-12-02 10:36:58.897658832 +0000 UTC m=+1072.227797712" watchObservedRunningTime="2025-12-02 10:36:58.904150869 +0000 UTC m=+1072.234289729" Dec 02 10:36:58 crc kubenswrapper[4679]: I1202 10:36:58.917551 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 02 10:36:59 crc kubenswrapper[4679]: I1202 10:36:59.848765 4679 generic.go:334] "Generic (PLEG): container finished" podID="9a9d4b4d-2f9c-4f27-9f64-98efad54fded" containerID="70b4b33f544f64dc764e7ec9caa370bfb11b5738c33f8ad8201a22eb12e11b67" exitCode=0 Dec 02 10:36:59 crc kubenswrapper[4679]: I1202 10:36:59.848872 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z6rx5" event={"ID":"9a9d4b4d-2f9c-4f27-9f64-98efad54fded","Type":"ContainerDied","Data":"70b4b33f544f64dc764e7ec9caa370bfb11b5738c33f8ad8201a22eb12e11b67"} Dec 02 10:36:59 crc kubenswrapper[4679]: I1202 10:36:59.849170 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z6rx5" event={"ID":"9a9d4b4d-2f9c-4f27-9f64-98efad54fded","Type":"ContainerStarted","Data":"e9ef242fd88087f5e376467438cf151786ba6b36053e1963f803d5f8d687d66b"} Dec 02 10:36:59 crc kubenswrapper[4679]: I1202 10:36:59.849190 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z6rx5" event={"ID":"9a9d4b4d-2f9c-4f27-9f64-98efad54fded","Type":"ContainerStarted","Data":"7217d5ee269d8b2fada1e9c368defacdb28e2ad5988933c87813a87a346edf43"} Dec 02 10:37:00 crc kubenswrapper[4679]: I1202 10:37:00.866943 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2a75d52-ee01-460e-b15b-70bff6593ef1","Type":"ContainerStarted","Data":"44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b"} Dec 02 10:37:00 crc kubenswrapper[4679]: I1202 10:37:00.867832 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 02 10:37:00 crc kubenswrapper[4679]: I1202 10:37:00.867851 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:37:00 crc kubenswrapper[4679]: I1202 10:37:00.867862 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:37:00 crc kubenswrapper[4679]: I1202 10:37:00.904455 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-z6rx5" podStartSLOduration=24.973293685 podStartE2EDuration="27.904433433s" podCreationTimestamp="2025-12-02 10:36:33 +0000 UTC" firstStartedPulling="2025-12-02 10:36:54.560124274 +0000 UTC m=+1067.890263134" lastFinishedPulling="2025-12-02 10:36:57.491264022 +0000 UTC m=+1070.821402882" observedRunningTime="2025-12-02 10:37:00.903521857 +0000 UTC m=+1074.233660727" watchObservedRunningTime="2025-12-02 10:37:00.904433433 +0000 UTC m=+1074.234572293" Dec 02 10:37:00 crc kubenswrapper[4679]: I1202 10:37:00.933688 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=23.986864265 podStartE2EDuration="30.933669795s" podCreationTimestamp="2025-12-02 10:36:30 +0000 UTC" firstStartedPulling="2025-12-02 10:36:53.487145321 +0000 UTC m=+1066.817284181" lastFinishedPulling="2025-12-02 10:37:00.433950851 +0000 UTC m=+1073.764089711" observedRunningTime="2025-12-02 10:37:00.928208858 +0000 UTC m=+1074.258347718" watchObservedRunningTime="2025-12-02 10:37:00.933669795 +0000 UTC m=+1074.263808655" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.035273 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sv9tw"] Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.086840 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hdg6r"] Dec 02 10:37:01 crc kubenswrapper[4679]: E1202 10:37:01.087216 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="extract-utilities" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.087238 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="extract-utilities" Dec 02 10:37:01 crc kubenswrapper[4679]: E1202 10:37:01.087268 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="registry-server" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.087278 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="registry-server" Dec 02 10:37:01 crc kubenswrapper[4679]: E1202 10:37:01.087612 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="extract-content" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.087632 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="extract-content" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.087845 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="96633b7d-7894-4e3e-adf2-ce16371793f5" containerName="registry-server" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.088973 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.111464 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hdg6r"] Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.201857 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-config\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.201930 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.202157 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np96b\" (UniqueName: \"kubernetes.io/projected/5f2d2380-df2a-4425-a3d3-16238531efd3-kube-api-access-np96b\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.306775 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-config\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.307544 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.307728 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np96b\" (UniqueName: \"kubernetes.io/projected/5f2d2380-df2a-4425-a3d3-16238531efd3-kube-api-access-np96b\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.308389 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-config\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.309294 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.362551 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np96b\" (UniqueName: \"kubernetes.io/projected/5f2d2380-df2a-4425-a3d3-16238531efd3-kube-api-access-np96b\") pod \"dnsmasq-dns-7cb5889db5-hdg6r\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.419640 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.521560 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.611151 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-config\") pod \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.611805 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-config" (OuterVolumeSpecName: "config") pod "227e410d-8f47-41ac-9f78-be5cbeb6c16f" (UID: "227e410d-8f47-41ac-9f78-be5cbeb6c16f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.617591 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrx6b\" (UniqueName: \"kubernetes.io/projected/227e410d-8f47-41ac-9f78-be5cbeb6c16f-kube-api-access-jrx6b\") pod \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.617755 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-dns-svc\") pod \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\" (UID: \"227e410d-8f47-41ac-9f78-be5cbeb6c16f\") " Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.620581 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "227e410d-8f47-41ac-9f78-be5cbeb6c16f" (UID: "227e410d-8f47-41ac-9f78-be5cbeb6c16f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.621882 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.621914 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/227e410d-8f47-41ac-9f78-be5cbeb6c16f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.627292 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/227e410d-8f47-41ac-9f78-be5cbeb6c16f-kube-api-access-jrx6b" (OuterVolumeSpecName: "kube-api-access-jrx6b") pod "227e410d-8f47-41ac-9f78-be5cbeb6c16f" (UID: "227e410d-8f47-41ac-9f78-be5cbeb6c16f"). InnerVolumeSpecName "kube-api-access-jrx6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.723598 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrx6b\" (UniqueName: \"kubernetes.io/projected/227e410d-8f47-41ac-9f78-be5cbeb6c16f-kube-api-access-jrx6b\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.725748 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hdg6r"] Dec 02 10:37:01 crc kubenswrapper[4679]: W1202 10:37:01.737496 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f2d2380_df2a_4425_a3d3_16238531efd3.slice/crio-c11a8fc537332611792f819e1889b64c0fdbd08f7758afb34063c49c7d40a78e WatchSource:0}: Error finding container c11a8fc537332611792f819e1889b64c0fdbd08f7758afb34063c49c7d40a78e: Status 404 returned error can't find the container with id c11a8fc537332611792f819e1889b64c0fdbd08f7758afb34063c49c7d40a78e Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.888856 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" event={"ID":"227e410d-8f47-41ac-9f78-be5cbeb6c16f","Type":"ContainerDied","Data":"4a6b549897c34aa8a3162b6c05eb43daa63b1eb3c8f876b7145d86b9c9d1e93b"} Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.888958 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-sv9tw" Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.901371 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" event={"ID":"5f2d2380-df2a-4425-a3d3-16238531efd3","Type":"ContainerStarted","Data":"c11a8fc537332611792f819e1889b64c0fdbd08f7758afb34063c49c7d40a78e"} Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.966269 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sv9tw"] Dec 02 10:37:01 crc kubenswrapper[4679]: I1202 10:37:01.989257 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-sv9tw"] Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.174026 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.179081 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.182542 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xjnzd" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.182733 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.184482 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.185268 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.207923 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.334337 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d8428a26-75d9-4d87-b31d-79329264e14e-cache\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.334426 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d8428a26-75d9-4d87-b31d-79329264e14e-lock\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.334955 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.335150 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwld2\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-kube-api-access-cwld2\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.335288 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.436790 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.436857 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwld2\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-kube-api-access-cwld2\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.436901 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.436929 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d8428a26-75d9-4d87-b31d-79329264e14e-cache\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.436956 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d8428a26-75d9-4d87-b31d-79329264e14e-lock\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.437512 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: E1202 10:37:02.437762 4679 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 10:37:02 crc kubenswrapper[4679]: E1202 10:37:02.437818 4679 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 10:37:02 crc kubenswrapper[4679]: E1202 10:37:02.437891 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift podName:d8428a26-75d9-4d87-b31d-79329264e14e nodeName:}" failed. No retries permitted until 2025-12-02 10:37:02.937862189 +0000 UTC m=+1076.268001049 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift") pod "swift-storage-0" (UID: "d8428a26-75d9-4d87-b31d-79329264e14e") : configmap "swift-ring-files" not found Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.438062 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d8428a26-75d9-4d87-b31d-79329264e14e-lock\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.438300 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d8428a26-75d9-4d87-b31d-79329264e14e-cache\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.461186 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwld2\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-kube-api-access-cwld2\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.463076 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.664625 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-2wk8m"] Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.665885 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.668161 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.668389 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.668833 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.687496 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2wk8m"] Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844233 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-ring-data-devices\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844332 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-scripts\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844408 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-combined-ca-bundle\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844435 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/714d01a2-8f79-41d8-a16a-47c35a69fe8b-etc-swift\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844461 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m44b9\" (UniqueName: \"kubernetes.io/projected/714d01a2-8f79-41d8-a16a-47c35a69fe8b-kube-api-access-m44b9\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844489 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-dispersionconf\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.844537 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-swiftconf\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.927160 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="227e410d-8f47-41ac-9f78-be5cbeb6c16f" path="/var/lib/kubelet/pods/227e410d-8f47-41ac-9f78-be5cbeb6c16f/volumes" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.929590 4679 generic.go:334] "Generic (PLEG): container finished" podID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerID="e33e703e5278e8dc82542a41a0c09cc2e15e17fddc7874b18591f711340d15ab" exitCode=0 Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.929631 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" event={"ID":"5f2d2380-df2a-4425-a3d3-16238531efd3","Type":"ContainerDied","Data":"e33e703e5278e8dc82542a41a0c09cc2e15e17fddc7874b18591f711340d15ab"} Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946485 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-ring-data-devices\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946573 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-scripts\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946730 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-combined-ca-bundle\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946758 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/714d01a2-8f79-41d8-a16a-47c35a69fe8b-etc-swift\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946784 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m44b9\" (UniqueName: \"kubernetes.io/projected/714d01a2-8f79-41d8-a16a-47c35a69fe8b-kube-api-access-m44b9\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946817 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-dispersionconf\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946845 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.946875 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-swiftconf\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: E1202 10:37:02.949212 4679 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 10:37:02 crc kubenswrapper[4679]: E1202 10:37:02.949251 4679 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 10:37:02 crc kubenswrapper[4679]: E1202 10:37:02.949331 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift podName:d8428a26-75d9-4d87-b31d-79329264e14e nodeName:}" failed. No retries permitted until 2025-12-02 10:37:03.94929504 +0000 UTC m=+1077.279433900 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift") pod "swift-storage-0" (UID: "d8428a26-75d9-4d87-b31d-79329264e14e") : configmap "swift-ring-files" not found Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.949460 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-scripts\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.949511 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-ring-data-devices\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.949823 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/714d01a2-8f79-41d8-a16a-47c35a69fe8b-etc-swift\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.953638 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-dispersionconf\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.953894 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-swiftconf\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.956412 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-combined-ca-bundle\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.967687 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m44b9\" (UniqueName: \"kubernetes.io/projected/714d01a2-8f79-41d8-a16a-47c35a69fe8b-kube-api-access-m44b9\") pod \"swift-ring-rebalance-2wk8m\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:02 crc kubenswrapper[4679]: I1202 10:37:02.986715 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:03 crc kubenswrapper[4679]: I1202 10:37:03.961898 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:03 crc kubenswrapper[4679]: E1202 10:37:03.962238 4679 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 10:37:03 crc kubenswrapper[4679]: E1202 10:37:03.963387 4679 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 10:37:03 crc kubenswrapper[4679]: E1202 10:37:03.963447 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift podName:d8428a26-75d9-4d87-b31d-79329264e14e nodeName:}" failed. No retries permitted until 2025-12-02 10:37:05.96342758 +0000 UTC m=+1079.293566440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift") pod "swift-storage-0" (UID: "d8428a26-75d9-4d87-b31d-79329264e14e") : configmap "swift-ring-files" not found Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.744470 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2wk8m"] Dec 02 10:37:04 crc kubenswrapper[4679]: W1202 10:37:04.749705 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod714d01a2_8f79_41d8_a16a_47c35a69fe8b.slice/crio-4be197bb1502648c45d479a6c472fc61bc46ebe7afe6b3edaf91e886155b4141 WatchSource:0}: Error finding container 4be197bb1502648c45d479a6c472fc61bc46ebe7afe6b3edaf91e886155b4141: Status 404 returned error can't find the container with id 4be197bb1502648c45d479a6c472fc61bc46ebe7afe6b3edaf91e886155b4141 Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.950351 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"285ae6d4-044a-4720-a495-8ca9eca431f3","Type":"ContainerStarted","Data":"e35c74a7300d2587eba516b99f97cdf145fb6d7014f9d14e79268305f4e5899c"} Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.952350 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2wk8m" event={"ID":"714d01a2-8f79-41d8-a16a-47c35a69fe8b","Type":"ContainerStarted","Data":"4be197bb1502648c45d479a6c472fc61bc46ebe7afe6b3edaf91e886155b4141"} Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.954778 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"71d9df94-9e39-409b-aa9e-e13a8f8f9ea2","Type":"ContainerStarted","Data":"610a5737c9c5f6cce79ba336b1dbe163a3a4b7490d5cac7c29a4b95dfb13a39a"} Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.957783 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" event={"ID":"5f2d2380-df2a-4425-a3d3-16238531efd3","Type":"ContainerStarted","Data":"628989f9e7aff7aa3c3ed0a4d7b9f5763065a4db5408848fb320a44d41e906c6"} Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.957920 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.959675 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e1f514bf-70bf-484c-b485-7e3d46a5c139","Type":"ContainerStarted","Data":"0949b2230ebffe9d53735ed881a0af6a2060d67899dedbeb3e490cf2c79022ed"} Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.981074 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=17.090112455 podStartE2EDuration="27.981053039s" podCreationTimestamp="2025-12-02 10:36:37 +0000 UTC" firstStartedPulling="2025-12-02 10:36:53.584689719 +0000 UTC m=+1066.914828569" lastFinishedPulling="2025-12-02 10:37:04.475630283 +0000 UTC m=+1077.805769153" observedRunningTime="2025-12-02 10:37:04.974169972 +0000 UTC m=+1078.304308832" watchObservedRunningTime="2025-12-02 10:37:04.981053039 +0000 UTC m=+1078.311191899" Dec 02 10:37:04 crc kubenswrapper[4679]: I1202 10:37:04.996720 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" podStartSLOduration=3.495638828 podStartE2EDuration="3.99669947s" podCreationTimestamp="2025-12-02 10:37:01 +0000 UTC" firstStartedPulling="2025-12-02 10:37:01.740051254 +0000 UTC m=+1075.070190114" lastFinishedPulling="2025-12-02 10:37:02.241111896 +0000 UTC m=+1075.571250756" observedRunningTime="2025-12-02 10:37:04.995462314 +0000 UTC m=+1078.325601224" watchObservedRunningTime="2025-12-02 10:37:04.99669947 +0000 UTC m=+1078.326838340" Dec 02 10:37:05 crc kubenswrapper[4679]: I1202 10:37:05.048025 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.954786911 podStartE2EDuration="32.048006617s" podCreationTimestamp="2025-12-02 10:36:33 +0000 UTC" firstStartedPulling="2025-12-02 10:36:53.433827837 +0000 UTC m=+1066.763966697" lastFinishedPulling="2025-12-02 10:37:04.527047543 +0000 UTC m=+1077.857186403" observedRunningTime="2025-12-02 10:37:05.044018462 +0000 UTC m=+1078.374157322" watchObservedRunningTime="2025-12-02 10:37:05.048006617 +0000 UTC m=+1078.378145477" Dec 02 10:37:05 crc kubenswrapper[4679]: I1202 10:37:05.611492 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 02 10:37:05 crc kubenswrapper[4679]: I1202 10:37:05.668782 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 02 10:37:05 crc kubenswrapper[4679]: I1202 10:37:05.970719 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae","Type":"ContainerStarted","Data":"b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f"} Dec 02 10:37:05 crc kubenswrapper[4679]: I1202 10:37:05.971803 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 02 10:37:05 crc kubenswrapper[4679]: I1202 10:37:05.997176 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:05 crc kubenswrapper[4679]: E1202 10:37:05.997917 4679 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 10:37:05 crc kubenswrapper[4679]: E1202 10:37:05.998022 4679 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 10:37:05 crc kubenswrapper[4679]: E1202 10:37:05.998070 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift podName:d8428a26-75d9-4d87-b31d-79329264e14e nodeName:}" failed. No retries permitted until 2025-12-02 10:37:09.998053523 +0000 UTC m=+1083.328192383 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift") pod "swift-storage-0" (UID: "d8428a26-75d9-4d87-b31d-79329264e14e") : configmap "swift-ring-files" not found Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.017996 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 02 10:37:06 crc kubenswrapper[4679]: E1202 10:37:06.143067 4679 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 02 10:37:06 crc kubenswrapper[4679]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/f27f0728-a9c2-4ce6-85ea-dab897da5f35/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 02 10:37:06 crc kubenswrapper[4679]: > podSandboxID="4be41f6f21d92f148243e348473862465ab22d915cea586d6da903cc2cce810c" Dec 02 10:37:06 crc kubenswrapper[4679]: E1202 10:37:06.143218 4679 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 02 10:37:06 crc kubenswrapper[4679]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ghhjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-zj2hk_openstack(f27f0728-a9c2-4ce6-85ea-dab897da5f35): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/f27f0728-a9c2-4ce6-85ea-dab897da5f35/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 02 10:37:06 crc kubenswrapper[4679]: > logger="UnhandledError" Dec 02 10:37:06 crc kubenswrapper[4679]: E1202 10:37:06.144405 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/f27f0728-a9c2-4ce6-85ea-dab897da5f35/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" podUID="f27f0728-a9c2-4ce6-85ea-dab897da5f35" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.260823 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zj2hk"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.292974 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-hjjnq"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.294544 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.301046 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.303735 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-hjjnq"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.332815 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kg24f"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.333918 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.337586 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.354212 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kg24f"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.402967 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j9gx\" (UniqueName: \"kubernetes.io/projected/142f095e-2fa0-478d-a611-50aad287d2c6-kube-api-access-7j9gx\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403024 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/142f095e-2fa0-478d-a611-50aad287d2c6-config\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403056 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j26xd\" (UniqueName: \"kubernetes.io/projected/e5f3e327-584f-420f-83d1-f8ac74cafe6f-kube-api-access-j26xd\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403089 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/142f095e-2fa0-478d-a611-50aad287d2c6-ovn-rundir\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403137 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403162 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f095e-2fa0-478d-a611-50aad287d2c6-combined-ca-bundle\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403179 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f095e-2fa0-478d-a611-50aad287d2c6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403384 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/142f095e-2fa0-478d-a611-50aad287d2c6-ovs-rundir\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403464 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.403486 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-config\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.475322 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hdg6r"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.502897 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmlpw"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.504477 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505148 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/142f095e-2fa0-478d-a611-50aad287d2c6-config\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505213 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j26xd\" (UniqueName: \"kubernetes.io/projected/e5f3e327-584f-420f-83d1-f8ac74cafe6f-kube-api-access-j26xd\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505253 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/142f095e-2fa0-478d-a611-50aad287d2c6-ovn-rundir\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505298 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505334 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f095e-2fa0-478d-a611-50aad287d2c6-combined-ca-bundle\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505352 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f095e-2fa0-478d-a611-50aad287d2c6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505371 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/142f095e-2fa0-478d-a611-50aad287d2c6-ovs-rundir\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505394 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505426 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-config\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505472 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j9gx\" (UniqueName: \"kubernetes.io/projected/142f095e-2fa0-478d-a611-50aad287d2c6-kube-api-access-7j9gx\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.505940 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/142f095e-2fa0-478d-a611-50aad287d2c6-config\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.506104 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/142f095e-2fa0-478d-a611-50aad287d2c6-ovn-rundir\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.506155 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/142f095e-2fa0-478d-a611-50aad287d2c6-ovs-rundir\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.506715 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.506904 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.507570 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.507620 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-config\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.516448 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/142f095e-2fa0-478d-a611-50aad287d2c6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.519070 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmlpw"] Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.526133 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142f095e-2fa0-478d-a611-50aad287d2c6-combined-ca-bundle\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.529714 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j9gx\" (UniqueName: \"kubernetes.io/projected/142f095e-2fa0-478d-a611-50aad287d2c6-kube-api-access-7j9gx\") pod \"ovn-controller-metrics-kg24f\" (UID: \"142f095e-2fa0-478d-a611-50aad287d2c6\") " pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.530329 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j26xd\" (UniqueName: \"kubernetes.io/projected/e5f3e327-584f-420f-83d1-f8ac74cafe6f-kube-api-access-j26xd\") pod \"dnsmasq-dns-6c89d5d749-hjjnq\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.606900 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.607021 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.607048 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6dm2\" (UniqueName: \"kubernetes.io/projected/08d8ff70-7182-405f-b253-96f212f13410-kube-api-access-c6dm2\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.607273 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-config\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.607342 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-dns-svc\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.622328 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.654843 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kg24f" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.716297 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-config\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.716427 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-dns-svc\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.716502 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.716558 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.716584 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6dm2\" (UniqueName: \"kubernetes.io/projected/08d8ff70-7182-405f-b253-96f212f13410-kube-api-access-c6dm2\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.718283 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-config\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.718373 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.718944 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.719041 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-dns-svc\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.766598 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6dm2\" (UniqueName: \"kubernetes.io/projected/08d8ff70-7182-405f-b253-96f212f13410-kube-api-access-c6dm2\") pod \"dnsmasq-dns-698758b865-qmlpw\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.905857 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:06 crc kubenswrapper[4679]: I1202 10:37:06.979169 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="dnsmasq-dns" containerID="cri-o://628989f9e7aff7aa3c3ed0a4d7b9f5763065a4db5408848fb320a44d41e906c6" gracePeriod=10 Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.178168 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-hjjnq"] Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.278526 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.293693 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kg24f"] Dec 02 10:37:07 crc kubenswrapper[4679]: W1202 10:37:07.315929 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod142f095e_2fa0_478d_a611_50aad287d2c6.slice/crio-bdd83915a93f452386237cb5c6c7930f9c804779c07bacc2e4d8e60d841e58e3 WatchSource:0}: Error finding container bdd83915a93f452386237cb5c6c7930f9c804779c07bacc2e4d8e60d841e58e3: Status 404 returned error can't find the container with id bdd83915a93f452386237cb5c6c7930f9c804779c07bacc2e4d8e60d841e58e3 Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.338097 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-config\") pod \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.338177 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghhjk\" (UniqueName: \"kubernetes.io/projected/f27f0728-a9c2-4ce6-85ea-dab897da5f35-kube-api-access-ghhjk\") pod \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.338343 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-dns-svc\") pod \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\" (UID: \"f27f0728-a9c2-4ce6-85ea-dab897da5f35\") " Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.354740 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27f0728-a9c2-4ce6-85ea-dab897da5f35-kube-api-access-ghhjk" (OuterVolumeSpecName: "kube-api-access-ghhjk") pod "f27f0728-a9c2-4ce6-85ea-dab897da5f35" (UID: "f27f0728-a9c2-4ce6-85ea-dab897da5f35"). InnerVolumeSpecName "kube-api-access-ghhjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.376821 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f27f0728-a9c2-4ce6-85ea-dab897da5f35" (UID: "f27f0728-a9c2-4ce6-85ea-dab897da5f35"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.388050 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-config" (OuterVolumeSpecName: "config") pod "f27f0728-a9c2-4ce6-85ea-dab897da5f35" (UID: "f27f0728-a9c2-4ce6-85ea-dab897da5f35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.440649 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.440684 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghhjk\" (UniqueName: \"kubernetes.io/projected/f27f0728-a9c2-4ce6-85ea-dab897da5f35-kube-api-access-ghhjk\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.440701 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f27f0728-a9c2-4ce6-85ea-dab897da5f35-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.452642 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmlpw"] Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.845676 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.899944 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.991354 4679 generic.go:334] "Generic (PLEG): container finished" podID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerID="5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a" exitCode=0 Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.991473 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" event={"ID":"e5f3e327-584f-420f-83d1-f8ac74cafe6f","Type":"ContainerDied","Data":"5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a"} Dec 02 10:37:07 crc kubenswrapper[4679]: I1202 10:37:07.991549 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" event={"ID":"e5f3e327-584f-420f-83d1-f8ac74cafe6f","Type":"ContainerStarted","Data":"4a4683011cca9c7a636bcbfc3320177121018a64429957e840a59f2d51a017a8"} Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.001772 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.005294 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zj2hk" event={"ID":"f27f0728-a9c2-4ce6-85ea-dab897da5f35","Type":"ContainerDied","Data":"4be41f6f21d92f148243e348473862465ab22d915cea586d6da903cc2cce810c"} Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.010217 4679 generic.go:334] "Generic (PLEG): container finished" podID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerID="628989f9e7aff7aa3c3ed0a4d7b9f5763065a4db5408848fb320a44d41e906c6" exitCode=0 Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.010350 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" event={"ID":"5f2d2380-df2a-4425-a3d3-16238531efd3","Type":"ContainerDied","Data":"628989f9e7aff7aa3c3ed0a4d7b9f5763065a4db5408848fb320a44d41e906c6"} Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.015027 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kg24f" event={"ID":"142f095e-2fa0-478d-a611-50aad287d2c6","Type":"ContainerStarted","Data":"bdd83915a93f452386237cb5c6c7930f9c804779c07bacc2e4d8e60d841e58e3"} Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.015460 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.061265 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.079125 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zj2hk"] Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.086136 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zj2hk"] Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.229366 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.231889 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.234799 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.234896 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-lmbxx" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.235030 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.235063 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.265972 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2e8d7b-ca00-404b-a11b-4ea010116548-scripts\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.266034 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.266162 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2e8d7b-ca00-404b-a11b-4ea010116548-config\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.266184 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae2e8d7b-ca00-404b-a11b-4ea010116548-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.266208 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.266262 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.266281 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6pfb\" (UniqueName: \"kubernetes.io/projected/ae2e8d7b-ca00-404b-a11b-4ea010116548-kube-api-access-t6pfb\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.267373 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368041 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6pfb\" (UniqueName: \"kubernetes.io/projected/ae2e8d7b-ca00-404b-a11b-4ea010116548-kube-api-access-t6pfb\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368086 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368147 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2e8d7b-ca00-404b-a11b-4ea010116548-scripts\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368176 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368219 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2e8d7b-ca00-404b-a11b-4ea010116548-config\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368237 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae2e8d7b-ca00-404b-a11b-4ea010116548-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.368255 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.369420 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2e8d7b-ca00-404b-a11b-4ea010116548-scripts\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.369531 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2e8d7b-ca00-404b-a11b-4ea010116548-config\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.370387 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae2e8d7b-ca00-404b-a11b-4ea010116548-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.374608 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.374625 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.374658 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2e8d7b-ca00-404b-a11b-4ea010116548-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.385656 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6pfb\" (UniqueName: \"kubernetes.io/projected/ae2e8d7b-ca00-404b-a11b-4ea010116548-kube-api-access-t6pfb\") pod \"ovn-northd-0\" (UID: \"ae2e8d7b-ca00-404b-a11b-4ea010116548\") " pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.536350 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.536413 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.571673 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.616932 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 02 10:37:08 crc kubenswrapper[4679]: I1202 10:37:08.917972 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f27f0728-a9c2-4ce6-85ea-dab897da5f35" path="/var/lib/kubelet/pods/f27f0728-a9c2-4ce6-85ea-dab897da5f35/volumes" Dec 02 10:37:09 crc kubenswrapper[4679]: I1202 10:37:09.023731 4679 generic.go:334] "Generic (PLEG): container finished" podID="e1f514bf-70bf-484c-b485-7e3d46a5c139" containerID="0949b2230ebffe9d53735ed881a0af6a2060d67899dedbeb3e490cf2c79022ed" exitCode=0 Dec 02 10:37:09 crc kubenswrapper[4679]: I1202 10:37:09.023803 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e1f514bf-70bf-484c-b485-7e3d46a5c139","Type":"ContainerDied","Data":"0949b2230ebffe9d53735ed881a0af6a2060d67899dedbeb3e490cf2c79022ed"} Dec 02 10:37:09 crc kubenswrapper[4679]: I1202 10:37:09.027232 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"416d04f1-7afb-4bdf-a5d5-1a0c381513f9","Type":"ContainerStarted","Data":"df9a6c6da390576c9b707919a8e3e5b6cf8c75ee5cff628082a007e62d19b02d"} Dec 02 10:37:09 crc kubenswrapper[4679]: I1202 10:37:09.120300 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 02 10:37:10 crc kubenswrapper[4679]: I1202 10:37:10.099733 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:10 crc kubenswrapper[4679]: E1202 10:37:10.099976 4679 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 10:37:10 crc kubenswrapper[4679]: E1202 10:37:10.100014 4679 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 10:37:10 crc kubenswrapper[4679]: E1202 10:37:10.100099 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift podName:d8428a26-75d9-4d87-b31d-79329264e14e nodeName:}" failed. No retries permitted until 2025-12-02 10:37:18.100072791 +0000 UTC m=+1091.430211651 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift") pod "swift-storage-0" (UID: "d8428a26-75d9-4d87-b31d-79329264e14e") : configmap "swift-ring-files" not found Dec 02 10:37:10 crc kubenswrapper[4679]: I1202 10:37:10.947793 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.073858 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmlpw" event={"ID":"08d8ff70-7182-405f-b253-96f212f13410","Type":"ContainerStarted","Data":"f76cec11eed3aac7e769da7a5e6b1cb25b72e9a53fa477d49ca227359625f32a"} Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.077195 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" event={"ID":"5f2d2380-df2a-4425-a3d3-16238531efd3","Type":"ContainerDied","Data":"c11a8fc537332611792f819e1889b64c0fdbd08f7758afb34063c49c7d40a78e"} Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.077379 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c11a8fc537332611792f819e1889b64c0fdbd08f7758afb34063c49c7d40a78e" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.240032 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.354580 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-config\") pod \"5f2d2380-df2a-4425-a3d3-16238531efd3\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.355080 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-dns-svc\") pod \"5f2d2380-df2a-4425-a3d3-16238531efd3\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.355119 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np96b\" (UniqueName: \"kubernetes.io/projected/5f2d2380-df2a-4425-a3d3-16238531efd3-kube-api-access-np96b\") pod \"5f2d2380-df2a-4425-a3d3-16238531efd3\" (UID: \"5f2d2380-df2a-4425-a3d3-16238531efd3\") " Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.358744 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f2d2380-df2a-4425-a3d3-16238531efd3-kube-api-access-np96b" (OuterVolumeSpecName: "kube-api-access-np96b") pod "5f2d2380-df2a-4425-a3d3-16238531efd3" (UID: "5f2d2380-df2a-4425-a3d3-16238531efd3"). InnerVolumeSpecName "kube-api-access-np96b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.391367 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-config" (OuterVolumeSpecName: "config") pod "5f2d2380-df2a-4425-a3d3-16238531efd3" (UID: "5f2d2380-df2a-4425-a3d3-16238531efd3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.394552 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f2d2380-df2a-4425-a3d3-16238531efd3" (UID: "5f2d2380-df2a-4425-a3d3-16238531efd3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.458051 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.458086 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np96b\" (UniqueName: \"kubernetes.io/projected/5f2d2380-df2a-4425-a3d3-16238531efd3-kube-api-access-np96b\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.458099 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f2d2380-df2a-4425-a3d3-16238531efd3-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:13 crc kubenswrapper[4679]: I1202 10:37:13.480653 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 02 10:37:13 crc kubenswrapper[4679]: W1202 10:37:13.596939 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae2e8d7b_ca00_404b_a11b_4ea010116548.slice/crio-70c00b3075a7c49f8a522233e396aded1306f17a2d09422b741bdbffd73f4322 WatchSource:0}: Error finding container 70c00b3075a7c49f8a522233e396aded1306f17a2d09422b741bdbffd73f4322: Status 404 returned error can't find the container with id 70c00b3075a7c49f8a522233e396aded1306f17a2d09422b741bdbffd73f4322 Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.086079 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e1f514bf-70bf-484c-b485-7e3d46a5c139","Type":"ContainerStarted","Data":"7fa0c0e68492788d1d449a6e1695e47c1d86f7cf54b6eefccd1739025860784e"} Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.087261 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2wk8m" event={"ID":"714d01a2-8f79-41d8-a16a-47c35a69fe8b","Type":"ContainerStarted","Data":"a13f58517620f0377f8604183d138477d54d5988d109c3e7e3423f6ae4940118"} Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.090136 4679 generic.go:334] "Generic (PLEG): container finished" podID="08d8ff70-7182-405f-b253-96f212f13410" containerID="3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747" exitCode=0 Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.090231 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmlpw" event={"ID":"08d8ff70-7182-405f-b253-96f212f13410","Type":"ContainerDied","Data":"3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747"} Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.091601 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" event={"ID":"e5f3e327-584f-420f-83d1-f8ac74cafe6f","Type":"ContainerStarted","Data":"34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a"} Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.091733 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.106608 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ae2e8d7b-ca00-404b-a11b-4ea010116548","Type":"ContainerStarted","Data":"70c00b3075a7c49f8a522233e396aded1306f17a2d09422b741bdbffd73f4322"} Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.111538 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371987.743256 podStartE2EDuration="49.111520344s" podCreationTimestamp="2025-12-02 10:36:25 +0000 UTC" firstStartedPulling="2025-12-02 10:36:27.809369778 +0000 UTC m=+1041.139508648" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:14.104356597 +0000 UTC m=+1087.434495477" watchObservedRunningTime="2025-12-02 10:37:14.111520344 +0000 UTC m=+1087.441659204" Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.118279 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.118278 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kg24f" event={"ID":"142f095e-2fa0-478d-a611-50aad287d2c6","Type":"ContainerStarted","Data":"85fb97679c77878bf3336eadbf69bc2d932469d2056248ab34ae89a8a63c54b8"} Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.152566 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" podStartSLOduration=8.152546855 podStartE2EDuration="8.152546855s" podCreationTimestamp="2025-12-02 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:14.144282617 +0000 UTC m=+1087.474421477" watchObservedRunningTime="2025-12-02 10:37:14.152546855 +0000 UTC m=+1087.482685715" Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.170024 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-2wk8m" podStartSLOduration=3.363457568 podStartE2EDuration="12.170005947s" podCreationTimestamp="2025-12-02 10:37:02 +0000 UTC" firstStartedPulling="2025-12-02 10:37:04.754450689 +0000 UTC m=+1078.084589539" lastFinishedPulling="2025-12-02 10:37:13.560999058 +0000 UTC m=+1086.891137918" observedRunningTime="2025-12-02 10:37:14.16418646 +0000 UTC m=+1087.494325320" watchObservedRunningTime="2025-12-02 10:37:14.170005947 +0000 UTC m=+1087.500144807" Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.205127 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kg24f" podStartSLOduration=8.205103177 podStartE2EDuration="8.205103177s" podCreationTimestamp="2025-12-02 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:14.200937307 +0000 UTC m=+1087.531076167" watchObservedRunningTime="2025-12-02 10:37:14.205103177 +0000 UTC m=+1087.535242037" Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.229578 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hdg6r"] Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.240001 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-hdg6r"] Dec 02 10:37:14 crc kubenswrapper[4679]: I1202 10:37:14.922557 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" path="/var/lib/kubelet/pods/5f2d2380-df2a-4425-a3d3-16238531efd3/volumes" Dec 02 10:37:15 crc kubenswrapper[4679]: I1202 10:37:15.130737 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmlpw" event={"ID":"08d8ff70-7182-405f-b253-96f212f13410","Type":"ContainerStarted","Data":"563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b"} Dec 02 10:37:15 crc kubenswrapper[4679]: I1202 10:37:15.131535 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:15 crc kubenswrapper[4679]: I1202 10:37:15.161756 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qmlpw" podStartSLOduration=9.161736182 podStartE2EDuration="9.161736182s" podCreationTimestamp="2025-12-02 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:15.152822556 +0000 UTC m=+1088.482961426" watchObservedRunningTime="2025-12-02 10:37:15.161736182 +0000 UTC m=+1088.491875042" Dec 02 10:37:16 crc kubenswrapper[4679]: I1202 10:37:16.138714 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ae2e8d7b-ca00-404b-a11b-4ea010116548","Type":"ContainerStarted","Data":"5663d1d5c5f65850fb872baa20f1ff812eada81a9fd67a2d9e3bad170eca45c6"} Dec 02 10:37:16 crc kubenswrapper[4679]: I1202 10:37:16.421411 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-hdg6r" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.108:5353: i/o timeout" Dec 02 10:37:17 crc kubenswrapper[4679]: I1202 10:37:17.149095 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ae2e8d7b-ca00-404b-a11b-4ea010116548","Type":"ContainerStarted","Data":"996fbd19c68d879133169d188862554d8b95dab12e790c9d51a6bcdcb998832c"} Dec 02 10:37:17 crc kubenswrapper[4679]: I1202 10:37:17.257071 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 02 10:37:17 crc kubenswrapper[4679]: I1202 10:37:17.257153 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 02 10:37:18 crc kubenswrapper[4679]: I1202 10:37:18.152888 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:18 crc kubenswrapper[4679]: E1202 10:37:18.153157 4679 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 10:37:18 crc kubenswrapper[4679]: E1202 10:37:18.153338 4679 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 10:37:18 crc kubenswrapper[4679]: E1202 10:37:18.153397 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift podName:d8428a26-75d9-4d87-b31d-79329264e14e nodeName:}" failed. No retries permitted until 2025-12-02 10:37:34.153380171 +0000 UTC m=+1107.483519031 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift") pod "swift-storage-0" (UID: "d8428a26-75d9-4d87-b31d-79329264e14e") : configmap "swift-ring-files" not found Dec 02 10:37:18 crc kubenswrapper[4679]: I1202 10:37:18.155481 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 02 10:37:18 crc kubenswrapper[4679]: I1202 10:37:18.180726 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=8.71744634 podStartE2EDuration="10.180707477s" podCreationTimestamp="2025-12-02 10:37:08 +0000 UTC" firstStartedPulling="2025-12-02 10:37:13.599597459 +0000 UTC m=+1086.929736319" lastFinishedPulling="2025-12-02 10:37:15.062858546 +0000 UTC m=+1088.392997456" observedRunningTime="2025-12-02 10:37:18.172879692 +0000 UTC m=+1091.503018552" watchObservedRunningTime="2025-12-02 10:37:18.180707477 +0000 UTC m=+1091.510846337" Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.178740 4679 generic.go:334] "Generic (PLEG): container finished" podID="714d01a2-8f79-41d8-a16a-47c35a69fe8b" containerID="a13f58517620f0377f8604183d138477d54d5988d109c3e7e3423f6ae4940118" exitCode=0 Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.178819 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2wk8m" event={"ID":"714d01a2-8f79-41d8-a16a-47c35a69fe8b","Type":"ContainerDied","Data":"a13f58517620f0377f8604183d138477d54d5988d109c3e7e3423f6ae4940118"} Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.328224 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.393483 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.623502 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.912509 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:37:21 crc kubenswrapper[4679]: I1202 10:37:21.991593 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-hjjnq"] Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.185999 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerName="dnsmasq-dns" containerID="cri-o://34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a" gracePeriod=10 Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.664372 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.813053 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.847841 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-dispersionconf\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.847893 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-combined-ca-bundle\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.847966 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-scripts\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.847999 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-ring-data-devices\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.848049 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m44b9\" (UniqueName: \"kubernetes.io/projected/714d01a2-8f79-41d8-a16a-47c35a69fe8b-kube-api-access-m44b9\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.848148 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-swiftconf\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.848177 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/714d01a2-8f79-41d8-a16a-47c35a69fe8b-etc-swift\") pod \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\" (UID: \"714d01a2-8f79-41d8-a16a-47c35a69fe8b\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.848995 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.849374 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/714d01a2-8f79-41d8-a16a-47c35a69fe8b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.867472 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/714d01a2-8f79-41d8-a16a-47c35a69fe8b-kube-api-access-m44b9" (OuterVolumeSpecName: "kube-api-access-m44b9") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "kube-api-access-m44b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.870659 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.888548 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-scripts" (OuterVolumeSpecName: "scripts") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.894662 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.920504 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "714d01a2-8f79-41d8-a16a-47c35a69fe8b" (UID: "714d01a2-8f79-41d8-a16a-47c35a69fe8b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.949702 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j26xd\" (UniqueName: \"kubernetes.io/projected/e5f3e327-584f-420f-83d1-f8ac74cafe6f-kube-api-access-j26xd\") pod \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.949785 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-dns-svc\") pod \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.949955 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-ovsdbserver-sb\") pod \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950027 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-config\") pod \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\" (UID: \"e5f3e327-584f-420f-83d1-f8ac74cafe6f\") " Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950389 4679 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950410 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m44b9\" (UniqueName: \"kubernetes.io/projected/714d01a2-8f79-41d8-a16a-47c35a69fe8b-kube-api-access-m44b9\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950421 4679 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950429 4679 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/714d01a2-8f79-41d8-a16a-47c35a69fe8b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950437 4679 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950445 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714d01a2-8f79-41d8-a16a-47c35a69fe8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.950453 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/714d01a2-8f79-41d8-a16a-47c35a69fe8b-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.953342 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f3e327-584f-420f-83d1-f8ac74cafe6f-kube-api-access-j26xd" (OuterVolumeSpecName: "kube-api-access-j26xd") pod "e5f3e327-584f-420f-83d1-f8ac74cafe6f" (UID: "e5f3e327-584f-420f-83d1-f8ac74cafe6f"). InnerVolumeSpecName "kube-api-access-j26xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.992217 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e5f3e327-584f-420f-83d1-f8ac74cafe6f" (UID: "e5f3e327-584f-420f-83d1-f8ac74cafe6f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.992546 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-config" (OuterVolumeSpecName: "config") pod "e5f3e327-584f-420f-83d1-f8ac74cafe6f" (UID: "e5f3e327-584f-420f-83d1-f8ac74cafe6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:22 crc kubenswrapper[4679]: I1202 10:37:22.992770 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5f3e327-584f-420f-83d1-f8ac74cafe6f" (UID: "e5f3e327-584f-420f-83d1-f8ac74cafe6f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.052603 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.053036 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j26xd\" (UniqueName: \"kubernetes.io/projected/e5f3e327-584f-420f-83d1-f8ac74cafe6f-kube-api-access-j26xd\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.053085 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.053098 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5f3e327-584f-420f-83d1-f8ac74cafe6f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.194647 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-2wk8m" event={"ID":"714d01a2-8f79-41d8-a16a-47c35a69fe8b","Type":"ContainerDied","Data":"4be197bb1502648c45d479a6c472fc61bc46ebe7afe6b3edaf91e886155b4141"} Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.194684 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4be197bb1502648c45d479a6c472fc61bc46ebe7afe6b3edaf91e886155b4141" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.194696 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2wk8m" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.198115 4679 generic.go:334] "Generic (PLEG): container finished" podID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerID="34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a" exitCode=0 Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.198166 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" event={"ID":"e5f3e327-584f-420f-83d1-f8ac74cafe6f","Type":"ContainerDied","Data":"34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a"} Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.198178 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.198204 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-hjjnq" event={"ID":"e5f3e327-584f-420f-83d1-f8ac74cafe6f","Type":"ContainerDied","Data":"4a4683011cca9c7a636bcbfc3320177121018a64429957e840a59f2d51a017a8"} Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.198226 4679 scope.go:117] "RemoveContainer" containerID="34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.224837 4679 scope.go:117] "RemoveContainer" containerID="5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.235367 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-hjjnq"] Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.242699 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-hjjnq"] Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.257938 4679 scope.go:117] "RemoveContainer" containerID="34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a" Dec 02 10:37:23 crc kubenswrapper[4679]: E1202 10:37:23.258567 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a\": container with ID starting with 34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a not found: ID does not exist" containerID="34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.258608 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a"} err="failed to get container status \"34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a\": rpc error: code = NotFound desc = could not find container \"34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a\": container with ID starting with 34c125b9ad1a67cda59668a2ce3c45097a0e1667505026f805fdeb8bf58f977a not found: ID does not exist" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.258636 4679 scope.go:117] "RemoveContainer" containerID="5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a" Dec 02 10:37:23 crc kubenswrapper[4679]: E1202 10:37:23.258986 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a\": container with ID starting with 5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a not found: ID does not exist" containerID="5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a" Dec 02 10:37:23 crc kubenswrapper[4679]: I1202 10:37:23.259023 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a"} err="failed to get container status \"5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a\": rpc error: code = NotFound desc = could not find container \"5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a\": container with ID starting with 5029a5347f8e4d76048a29584611666f78b2d0c665986642220281760b4c8c0a not found: ID does not exist" Dec 02 10:37:24 crc kubenswrapper[4679]: I1202 10:37:24.918436 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" path="/var/lib/kubelet/pods/e5f3e327-584f-420f-83d1-f8ac74cafe6f/volumes" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.551869 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e4cf-account-create-update-fp8rh"] Dec 02 10:37:28 crc kubenswrapper[4679]: E1202 10:37:28.552401 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="dnsmasq-dns" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552417 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="dnsmasq-dns" Dec 02 10:37:28 crc kubenswrapper[4679]: E1202 10:37:28.552438 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="init" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552447 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="init" Dec 02 10:37:28 crc kubenswrapper[4679]: E1202 10:37:28.552474 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerName="init" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552483 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerName="init" Dec 02 10:37:28 crc kubenswrapper[4679]: E1202 10:37:28.552495 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="714d01a2-8f79-41d8-a16a-47c35a69fe8b" containerName="swift-ring-rebalance" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552505 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="714d01a2-8f79-41d8-a16a-47c35a69fe8b" containerName="swift-ring-rebalance" Dec 02 10:37:28 crc kubenswrapper[4679]: E1202 10:37:28.552530 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerName="dnsmasq-dns" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552538 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerName="dnsmasq-dns" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552720 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f3e327-584f-420f-83d1-f8ac74cafe6f" containerName="dnsmasq-dns" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552737 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="714d01a2-8f79-41d8-a16a-47c35a69fe8b" containerName="swift-ring-rebalance" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.552756 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f2d2380-df2a-4425-a3d3-16238531efd3" containerName="dnsmasq-dns" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.553477 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.556366 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.570957 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-w6dwk"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.575604 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.577984 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e4cf-account-create-update-fp8rh"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.584402 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-w6dwk"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.646700 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.649108 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2f792c0-0b54-4599-aea3-d5567072e6f7-operator-scripts\") pod \"keystone-e4cf-account-create-update-fp8rh\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.649153 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcngt\" (UniqueName: \"kubernetes.io/projected/e6f8df03-1845-4c2b-a30a-3d8186287f73-kube-api-access-mcngt\") pod \"keystone-db-create-w6dwk\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.649171 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6f8df03-1845-4c2b-a30a-3d8186287f73-operator-scripts\") pod \"keystone-db-create-w6dwk\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.649211 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjzkr\" (UniqueName: \"kubernetes.io/projected/c2f792c0-0b54-4599-aea3-d5567072e6f7-kube-api-access-hjzkr\") pod \"keystone-e4cf-account-create-update-fp8rh\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.750668 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2f792c0-0b54-4599-aea3-d5567072e6f7-operator-scripts\") pod \"keystone-e4cf-account-create-update-fp8rh\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.750715 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcngt\" (UniqueName: \"kubernetes.io/projected/e6f8df03-1845-4c2b-a30a-3d8186287f73-kube-api-access-mcngt\") pod \"keystone-db-create-w6dwk\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.750736 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6f8df03-1845-4c2b-a30a-3d8186287f73-operator-scripts\") pod \"keystone-db-create-w6dwk\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.750791 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjzkr\" (UniqueName: \"kubernetes.io/projected/c2f792c0-0b54-4599-aea3-d5567072e6f7-kube-api-access-hjzkr\") pod \"keystone-e4cf-account-create-update-fp8rh\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.751645 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2f792c0-0b54-4599-aea3-d5567072e6f7-operator-scripts\") pod \"keystone-e4cf-account-create-update-fp8rh\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.751816 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6f8df03-1845-4c2b-a30a-3d8186287f73-operator-scripts\") pod \"keystone-db-create-w6dwk\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.768551 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjzkr\" (UniqueName: \"kubernetes.io/projected/c2f792c0-0b54-4599-aea3-d5567072e6f7-kube-api-access-hjzkr\") pod \"keystone-e4cf-account-create-update-fp8rh\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.769561 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcngt\" (UniqueName: \"kubernetes.io/projected/e6f8df03-1845-4c2b-a30a-3d8186287f73-kube-api-access-mcngt\") pod \"keystone-db-create-w6dwk\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.826609 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-w84l8"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.827636 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.834001 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-w84l8"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.851941 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a6ecd7-a808-43a6-9562-953818cc9020-operator-scripts\") pod \"placement-db-create-w84l8\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.852450 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mqvl\" (UniqueName: \"kubernetes.io/projected/00a6ecd7-a808-43a6-9562-953818cc9020-kube-api-access-8mqvl\") pod \"placement-db-create-w84l8\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.878178 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.892920 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.934243 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bf39-account-create-update-9gvms"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.935263 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.938014 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.941106 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bf39-account-create-update-9gvms"] Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.964174 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mqvl\" (UniqueName: \"kubernetes.io/projected/00a6ecd7-a808-43a6-9562-953818cc9020-kube-api-access-8mqvl\") pod \"placement-db-create-w84l8\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.964336 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a6ecd7-a808-43a6-9562-953818cc9020-operator-scripts\") pod \"placement-db-create-w84l8\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.964997 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a6ecd7-a808-43a6-9562-953818cc9020-operator-scripts\") pod \"placement-db-create-w84l8\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.985170 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mqvl\" (UniqueName: \"kubernetes.io/projected/00a6ecd7-a808-43a6-9562-953818cc9020-kube-api-access-8mqvl\") pod \"placement-db-create-w84l8\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " pod="openstack/placement-db-create-w84l8" Dec 02 10:37:28 crc kubenswrapper[4679]: I1202 10:37:28.995446 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p5dsw" podUID="9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb" containerName="ovn-controller" probeResult="failure" output=< Dec 02 10:37:28 crc kubenswrapper[4679]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 02 10:37:28 crc kubenswrapper[4679]: > Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.065487 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fffms\" (UniqueName: \"kubernetes.io/projected/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-kube-api-access-fffms\") pod \"placement-bf39-account-create-update-9gvms\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.065638 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-operator-scripts\") pod \"placement-bf39-account-create-update-9gvms\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.133128 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-2xs5t"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.135603 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.142983 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2xs5t"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.154971 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w84l8" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.167502 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db8eb5a7-5fef-4f4a-905f-40103b0f513d-operator-scripts\") pod \"glance-db-create-2xs5t\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.167820 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5p4g\" (UniqueName: \"kubernetes.io/projected/db8eb5a7-5fef-4f4a-905f-40103b0f513d-kube-api-access-n5p4g\") pod \"glance-db-create-2xs5t\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.167980 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-operator-scripts\") pod \"placement-bf39-account-create-update-9gvms\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.168157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fffms\" (UniqueName: \"kubernetes.io/projected/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-kube-api-access-fffms\") pod \"placement-bf39-account-create-update-9gvms\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.169066 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-operator-scripts\") pod \"placement-bf39-account-create-update-9gvms\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.184942 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fffms\" (UniqueName: \"kubernetes.io/projected/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-kube-api-access-fffms\") pod \"placement-bf39-account-create-update-9gvms\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.242503 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-164a-account-create-update-bsvts"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.246756 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.249295 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.255892 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-164a-account-create-update-bsvts"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.273836 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db8eb5a7-5fef-4f4a-905f-40103b0f513d-operator-scripts\") pod \"glance-db-create-2xs5t\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.274439 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5p4g\" (UniqueName: \"kubernetes.io/projected/db8eb5a7-5fef-4f4a-905f-40103b0f513d-kube-api-access-n5p4g\") pod \"glance-db-create-2xs5t\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.274926 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db8eb5a7-5fef-4f4a-905f-40103b0f513d-operator-scripts\") pod \"glance-db-create-2xs5t\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.293477 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5p4g\" (UniqueName: \"kubernetes.io/projected/db8eb5a7-5fef-4f4a-905f-40103b0f513d-kube-api-access-n5p4g\") pod \"glance-db-create-2xs5t\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.370897 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.376869 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59eae751-1d2f-4ce3-a386-d2dcc0727717-operator-scripts\") pod \"glance-164a-account-create-update-bsvts\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.376952 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pbk5\" (UniqueName: \"kubernetes.io/projected/59eae751-1d2f-4ce3-a386-d2dcc0727717-kube-api-access-8pbk5\") pod \"glance-164a-account-create-update-bsvts\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.418731 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e4cf-account-create-update-fp8rh"] Dec 02 10:37:29 crc kubenswrapper[4679]: W1202 10:37:29.428234 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2f792c0_0b54_4599_aea3_d5567072e6f7.slice/crio-d79bd1cc45276bbbe513c494d077f08326eb92a502d0b55c4a8d47fb468c10f4 WatchSource:0}: Error finding container d79bd1cc45276bbbe513c494d077f08326eb92a502d0b55c4a8d47fb468c10f4: Status 404 returned error can't find the container with id d79bd1cc45276bbbe513c494d077f08326eb92a502d0b55c4a8d47fb468c10f4 Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.451934 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.478645 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pbk5\" (UniqueName: \"kubernetes.io/projected/59eae751-1d2f-4ce3-a386-d2dcc0727717-kube-api-access-8pbk5\") pod \"glance-164a-account-create-update-bsvts\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.478762 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59eae751-1d2f-4ce3-a386-d2dcc0727717-operator-scripts\") pod \"glance-164a-account-create-update-bsvts\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.479405 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59eae751-1d2f-4ce3-a386-d2dcc0727717-operator-scripts\") pod \"glance-164a-account-create-update-bsvts\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.496826 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pbk5\" (UniqueName: \"kubernetes.io/projected/59eae751-1d2f-4ce3-a386-d2dcc0727717-kube-api-access-8pbk5\") pod \"glance-164a-account-create-update-bsvts\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.536697 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-w6dwk"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.570255 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.685055 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-w84l8"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.833152 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bf39-account-create-update-9gvms"] Dec 02 10:37:29 crc kubenswrapper[4679]: I1202 10:37:29.914831 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2xs5t"] Dec 02 10:37:29 crc kubenswrapper[4679]: W1202 10:37:29.922607 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb8eb5a7_5fef_4f4a_905f_40103b0f513d.slice/crio-c0634c68c8e98f2d82e66fd30fafb166ea33a8e2e88b336072cdbd6bb4a43fa7 WatchSource:0}: Error finding container c0634c68c8e98f2d82e66fd30fafb166ea33a8e2e88b336072cdbd6bb4a43fa7: Status 404 returned error can't find the container with id c0634c68c8e98f2d82e66fd30fafb166ea33a8e2e88b336072cdbd6bb4a43fa7 Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.034261 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-164a-account-create-update-bsvts"] Dec 02 10:37:30 crc kubenswrapper[4679]: W1202 10:37:30.042478 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59eae751_1d2f_4ce3_a386_d2dcc0727717.slice/crio-668ff710ea936e4f69bdde67d4a98b218d8c82afbc5ea09badef2847b75bad95 WatchSource:0}: Error finding container 668ff710ea936e4f69bdde67d4a98b218d8c82afbc5ea09badef2847b75bad95: Status 404 returned error can't find the container with id 668ff710ea936e4f69bdde67d4a98b218d8c82afbc5ea09badef2847b75bad95 Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.263913 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-164a-account-create-update-bsvts" event={"ID":"59eae751-1d2f-4ce3-a386-d2dcc0727717","Type":"ContainerStarted","Data":"668ff710ea936e4f69bdde67d4a98b218d8c82afbc5ea09badef2847b75bad95"} Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.264980 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e4cf-account-create-update-fp8rh" event={"ID":"c2f792c0-0b54-4599-aea3-d5567072e6f7","Type":"ContainerStarted","Data":"d79bd1cc45276bbbe513c494d077f08326eb92a502d0b55c4a8d47fb468c10f4"} Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.267230 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w6dwk" event={"ID":"e6f8df03-1845-4c2b-a30a-3d8186287f73","Type":"ContainerStarted","Data":"41a255764934c57806da9b3b4577c3489a8aa91dc6700a5b7934f3216fc206d2"} Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.269386 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bf39-account-create-update-9gvms" event={"ID":"f5c4d5af-bdac-4486-b0ab-f389cd3331bb","Type":"ContainerStarted","Data":"044181bd3fb93163017fc824b261ec1f0e29f1bb66ce32c11ef1cb82015fc7f2"} Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.270579 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2xs5t" event={"ID":"db8eb5a7-5fef-4f4a-905f-40103b0f513d","Type":"ContainerStarted","Data":"c0634c68c8e98f2d82e66fd30fafb166ea33a8e2e88b336072cdbd6bb4a43fa7"} Dec 02 10:37:30 crc kubenswrapper[4679]: I1202 10:37:30.271614 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w84l8" event={"ID":"00a6ecd7-a808-43a6-9562-953818cc9020","Type":"ContainerStarted","Data":"9ac1ebd282cb365ee3103b8ddc424b00ef36597dabdd42e2395a349bf5a19535"} Dec 02 10:37:33 crc kubenswrapper[4679]: I1202 10:37:33.995150 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p5dsw" podUID="9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb" containerName="ovn-controller" probeResult="failure" output=< Dec 02 10:37:33 crc kubenswrapper[4679]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 02 10:37:33 crc kubenswrapper[4679]: > Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.015531 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.025087 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-z6rx5" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.161797 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.170384 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d8428a26-75d9-4d87-b31d-79329264e14e-etc-swift\") pod \"swift-storage-0\" (UID: \"d8428a26-75d9-4d87-b31d-79329264e14e\") " pod="openstack/swift-storage-0" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.260855 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p5dsw-config-4qbns"] Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.262374 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.267028 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p5dsw-config-4qbns"] Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.268852 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.307275 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.364345 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run-ovn\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.364394 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-additional-scripts\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.364450 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-scripts\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.364491 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-log-ovn\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.364507 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gglrw\" (UniqueName: \"kubernetes.io/projected/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-kube-api-access-gglrw\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.364565 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.472289 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.472393 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run-ovn\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.472423 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-additional-scripts\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.472489 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-scripts\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.472525 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-log-ovn\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.472547 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gglrw\" (UniqueName: \"kubernetes.io/projected/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-kube-api-access-gglrw\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.473596 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-additional-scripts\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.473805 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.473865 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run-ovn\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.473913 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-log-ovn\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.476287 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-scripts\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.506978 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gglrw\" (UniqueName: \"kubernetes.io/projected/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-kube-api-access-gglrw\") pod \"ovn-controller-p5dsw-config-4qbns\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.585951 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:34 crc kubenswrapper[4679]: I1202 10:37:34.906013 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 02 10:37:34 crc kubenswrapper[4679]: W1202 10:37:34.911533 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8428a26_75d9_4d87_b31d_79329264e14e.slice/crio-99893b6895a89de6e9d28681fd90ef813a0d12628aa4e87d10f89425bbfc0f53 WatchSource:0}: Error finding container 99893b6895a89de6e9d28681fd90ef813a0d12628aa4e87d10f89425bbfc0f53: Status 404 returned error can't find the container with id 99893b6895a89de6e9d28681fd90ef813a0d12628aa4e87d10f89425bbfc0f53 Dec 02 10:37:35 crc kubenswrapper[4679]: I1202 10:37:35.005684 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p5dsw-config-4qbns"] Dec 02 10:37:35 crc kubenswrapper[4679]: W1202 10:37:35.011616 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8bd6e13_5a1a_431a_b28a_08aee6a5f619.slice/crio-b44b58e7c45515cdc8ba0660e6a33d0b39a8ed828991982869586082b7cddd18 WatchSource:0}: Error finding container b44b58e7c45515cdc8ba0660e6a33d0b39a8ed828991982869586082b7cddd18: Status 404 returned error can't find the container with id b44b58e7c45515cdc8ba0660e6a33d0b39a8ed828991982869586082b7cddd18 Dec 02 10:37:35 crc kubenswrapper[4679]: I1202 10:37:35.320113 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p5dsw-config-4qbns" event={"ID":"b8bd6e13-5a1a-431a-b28a-08aee6a5f619","Type":"ContainerStarted","Data":"b44b58e7c45515cdc8ba0660e6a33d0b39a8ed828991982869586082b7cddd18"} Dec 02 10:37:35 crc kubenswrapper[4679]: I1202 10:37:35.321960 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"99893b6895a89de6e9d28681fd90ef813a0d12628aa4e87d10f89425bbfc0f53"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.349562 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w84l8" event={"ID":"00a6ecd7-a808-43a6-9562-953818cc9020","Type":"ContainerStarted","Data":"e3e3ffd41d2f57357f50ff15e951390ff9b6ef2feefc458118acf3b17dca7b95"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.352213 4679 generic.go:334] "Generic (PLEG): container finished" podID="b8bd6e13-5a1a-431a-b28a-08aee6a5f619" containerID="ddd15b0d3ba3c2197cf08f8a27e619215b6e7c5bd344af1a6861a46d7adca56c" exitCode=0 Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.352349 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p5dsw-config-4qbns" event={"ID":"b8bd6e13-5a1a-431a-b28a-08aee6a5f619","Type":"ContainerDied","Data":"ddd15b0d3ba3c2197cf08f8a27e619215b6e7c5bd344af1a6861a46d7adca56c"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.354409 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-164a-account-create-update-bsvts" event={"ID":"59eae751-1d2f-4ce3-a386-d2dcc0727717","Type":"ContainerStarted","Data":"8dae82f6710a97d71a8bb2dd5eb50004498520c0491b19bd114b3263258a730e"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.355866 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e4cf-account-create-update-fp8rh" event={"ID":"c2f792c0-0b54-4599-aea3-d5567072e6f7","Type":"ContainerStarted","Data":"2da94eb5efb39877d43b62507e84d1bea5bc435f7fbd004b02a008dc3d59e28a"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.357161 4679 generic.go:334] "Generic (PLEG): container finished" podID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerID="b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f" exitCode=0 Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.357263 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae","Type":"ContainerDied","Data":"b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.359154 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w6dwk" event={"ID":"e6f8df03-1845-4c2b-a30a-3d8186287f73","Type":"ContainerStarted","Data":"a8d697ea66b8450103078fcdf58a8daa4cb4f27754938a7c0c5db8e41bcec107"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.360929 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bf39-account-create-update-9gvms" event={"ID":"f5c4d5af-bdac-4486-b0ab-f389cd3331bb","Type":"ContainerStarted","Data":"cba53ad1d4dda8482a8887e03b1dd1b51a10ceaf76382f10be8b7df5b866dd4c"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.362371 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2xs5t" event={"ID":"db8eb5a7-5fef-4f4a-905f-40103b0f513d","Type":"ContainerStarted","Data":"c070b247a78fa29a7919fe681f74d9c5f9764b148e44474d917e3e58c3eb7bd0"} Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.370404 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-w84l8" podStartSLOduration=10.370387867 podStartE2EDuration="10.370387867s" podCreationTimestamp="2025-12-02 10:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:38.36356548 +0000 UTC m=+1111.693704370" watchObservedRunningTime="2025-12-02 10:37:38.370387867 +0000 UTC m=+1111.700526727" Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.386988 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-e4cf-account-create-update-fp8rh" podStartSLOduration=10.386973454 podStartE2EDuration="10.386973454s" podCreationTimestamp="2025-12-02 10:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:38.380566459 +0000 UTC m=+1111.710705329" watchObservedRunningTime="2025-12-02 10:37:38.386973454 +0000 UTC m=+1111.717112314" Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.422096 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-w6dwk" podStartSLOduration=10.422075224 podStartE2EDuration="10.422075224s" podCreationTimestamp="2025-12-02 10:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:38.420919651 +0000 UTC m=+1111.751058511" watchObservedRunningTime="2025-12-02 10:37:38.422075224 +0000 UTC m=+1111.752214084" Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.462823 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-2xs5t" podStartSLOduration=9.462802056 podStartE2EDuration="9.462802056s" podCreationTimestamp="2025-12-02 10:37:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:38.453413146 +0000 UTC m=+1111.783552016" watchObservedRunningTime="2025-12-02 10:37:38.462802056 +0000 UTC m=+1111.792940916" Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.476285 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-bf39-account-create-update-9gvms" podStartSLOduration=10.476266034 podStartE2EDuration="10.476266034s" podCreationTimestamp="2025-12-02 10:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:38.471231059 +0000 UTC m=+1111.801369929" watchObservedRunningTime="2025-12-02 10:37:38.476266034 +0000 UTC m=+1111.806404894" Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.486188 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-164a-account-create-update-bsvts" podStartSLOduration=9.486168619 podStartE2EDuration="9.486168619s" podCreationTimestamp="2025-12-02 10:37:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:38.483609455 +0000 UTC m=+1111.813748315" watchObservedRunningTime="2025-12-02 10:37:38.486168619 +0000 UTC m=+1111.816307479" Dec 02 10:37:38 crc kubenswrapper[4679]: I1202 10:37:38.992194 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-p5dsw" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.371820 4679 generic.go:334] "Generic (PLEG): container finished" podID="f5c4d5af-bdac-4486-b0ab-f389cd3331bb" containerID="cba53ad1d4dda8482a8887e03b1dd1b51a10ceaf76382f10be8b7df5b866dd4c" exitCode=0 Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.372293 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bf39-account-create-update-9gvms" event={"ID":"f5c4d5af-bdac-4486-b0ab-f389cd3331bb","Type":"ContainerDied","Data":"cba53ad1d4dda8482a8887e03b1dd1b51a10ceaf76382f10be8b7df5b866dd4c"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.375907 4679 generic.go:334] "Generic (PLEG): container finished" podID="db8eb5a7-5fef-4f4a-905f-40103b0f513d" containerID="c070b247a78fa29a7919fe681f74d9c5f9764b148e44474d917e3e58c3eb7bd0" exitCode=0 Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.375999 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2xs5t" event={"ID":"db8eb5a7-5fef-4f4a-905f-40103b0f513d","Type":"ContainerDied","Data":"c070b247a78fa29a7919fe681f74d9c5f9764b148e44474d917e3e58c3eb7bd0"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.377627 4679 generic.go:334] "Generic (PLEG): container finished" podID="00a6ecd7-a808-43a6-9562-953818cc9020" containerID="e3e3ffd41d2f57357f50ff15e951390ff9b6ef2feefc458118acf3b17dca7b95" exitCode=0 Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.377670 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w84l8" event={"ID":"00a6ecd7-a808-43a6-9562-953818cc9020","Type":"ContainerDied","Data":"e3e3ffd41d2f57357f50ff15e951390ff9b6ef2feefc458118acf3b17dca7b95"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.378964 4679 generic.go:334] "Generic (PLEG): container finished" podID="c2f792c0-0b54-4599-aea3-d5567072e6f7" containerID="2da94eb5efb39877d43b62507e84d1bea5bc435f7fbd004b02a008dc3d59e28a" exitCode=0 Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.379006 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e4cf-account-create-update-fp8rh" event={"ID":"c2f792c0-0b54-4599-aea3-d5567072e6f7","Type":"ContainerDied","Data":"2da94eb5efb39877d43b62507e84d1bea5bc435f7fbd004b02a008dc3d59e28a"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.380617 4679 generic.go:334] "Generic (PLEG): container finished" podID="59eae751-1d2f-4ce3-a386-d2dcc0727717" containerID="8dae82f6710a97d71a8bb2dd5eb50004498520c0491b19bd114b3263258a730e" exitCode=0 Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.380659 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-164a-account-create-update-bsvts" event={"ID":"59eae751-1d2f-4ce3-a386-d2dcc0727717","Type":"ContainerDied","Data":"8dae82f6710a97d71a8bb2dd5eb50004498520c0491b19bd114b3263258a730e"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.382360 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae","Type":"ContainerStarted","Data":"b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.382582 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.383852 4679 generic.go:334] "Generic (PLEG): container finished" podID="e6f8df03-1845-4c2b-a30a-3d8186287f73" containerID="a8d697ea66b8450103078fcdf58a8daa4cb4f27754938a7c0c5db8e41bcec107" exitCode=0 Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.383894 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w6dwk" event={"ID":"e6f8df03-1845-4c2b-a30a-3d8186287f73","Type":"ContainerDied","Data":"a8d697ea66b8450103078fcdf58a8daa4cb4f27754938a7c0c5db8e41bcec107"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.387486 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"6ec89f650512fb18e473b03a6fd1a39ec1158764e7bae79e3ac4e9b59b4dd12a"} Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.452341 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.118517295 podStartE2EDuration="1m15.452319648s" podCreationTimestamp="2025-12-02 10:36:24 +0000 UTC" firstStartedPulling="2025-12-02 10:36:26.14215823 +0000 UTC m=+1039.472297090" lastFinishedPulling="2025-12-02 10:37:04.475960583 +0000 UTC m=+1077.806099443" observedRunningTime="2025-12-02 10:37:39.445462421 +0000 UTC m=+1112.775601301" watchObservedRunningTime="2025-12-02 10:37:39.452319648 +0000 UTC m=+1112.782458508" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.697823 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.764525 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-log-ovn\") pod \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.764591 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b8bd6e13-5a1a-431a-b28a-08aee6a5f619" (UID: "b8bd6e13-5a1a-431a-b28a-08aee6a5f619"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.764757 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-additional-scripts\") pod \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.764811 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gglrw\" (UniqueName: \"kubernetes.io/projected/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-kube-api-access-gglrw\") pod \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.764838 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run-ovn\") pod \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.764919 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-scripts\") pod \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.765012 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b8bd6e13-5a1a-431a-b28a-08aee6a5f619" (UID: "b8bd6e13-5a1a-431a-b28a-08aee6a5f619"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.765546 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b8bd6e13-5a1a-431a-b28a-08aee6a5f619" (UID: "b8bd6e13-5a1a-431a-b28a-08aee6a5f619"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.765659 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run" (OuterVolumeSpecName: "var-run") pod "b8bd6e13-5a1a-431a-b28a-08aee6a5f619" (UID: "b8bd6e13-5a1a-431a-b28a-08aee6a5f619"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.766049 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-scripts" (OuterVolumeSpecName: "scripts") pod "b8bd6e13-5a1a-431a-b28a-08aee6a5f619" (UID: "b8bd6e13-5a1a-431a-b28a-08aee6a5f619"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.765095 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run\") pod \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\" (UID: \"b8bd6e13-5a1a-431a-b28a-08aee6a5f619\") " Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.766906 4679 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.766923 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.766932 4679 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-run\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.766940 4679 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.766952 4679 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.771942 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-kube-api-access-gglrw" (OuterVolumeSpecName: "kube-api-access-gglrw") pod "b8bd6e13-5a1a-431a-b28a-08aee6a5f619" (UID: "b8bd6e13-5a1a-431a-b28a-08aee6a5f619"). InnerVolumeSpecName "kube-api-access-gglrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:39 crc kubenswrapper[4679]: I1202 10:37:39.869427 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gglrw\" (UniqueName: \"kubernetes.io/projected/b8bd6e13-5a1a-431a-b28a-08aee6a5f619-kube-api-access-gglrw\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.402974 4679 generic.go:334] "Generic (PLEG): container finished" podID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerID="df9a6c6da390576c9b707919a8e3e5b6cf8c75ee5cff628082a007e62d19b02d" exitCode=0 Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.403065 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"416d04f1-7afb-4bdf-a5d5-1a0c381513f9","Type":"ContainerDied","Data":"df9a6c6da390576c9b707919a8e3e5b6cf8c75ee5cff628082a007e62d19b02d"} Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.409773 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p5dsw-config-4qbns" event={"ID":"b8bd6e13-5a1a-431a-b28a-08aee6a5f619","Type":"ContainerDied","Data":"b44b58e7c45515cdc8ba0660e6a33d0b39a8ed828991982869586082b7cddd18"} Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.409816 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p5dsw-config-4qbns" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.409835 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b44b58e7c45515cdc8ba0660e6a33d0b39a8ed828991982869586082b7cddd18" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.421474 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"5a4f6661ddaa3133bac0f56d3dd9e9792f3f8bab4a43ec2d82d09690aba921a8"} Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.421522 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"fe4ff5b5dc3b117f255b1a457c1cd950ac0279097bc29a58df888eb9d96ff3a6"} Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.772537 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w84l8" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.810370 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-p5dsw-config-4qbns"] Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.814202 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-p5dsw-config-4qbns"] Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.893124 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mqvl\" (UniqueName: \"kubernetes.io/projected/00a6ecd7-a808-43a6-9562-953818cc9020-kube-api-access-8mqvl\") pod \"00a6ecd7-a808-43a6-9562-953818cc9020\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.893227 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a6ecd7-a808-43a6-9562-953818cc9020-operator-scripts\") pod \"00a6ecd7-a808-43a6-9562-953818cc9020\" (UID: \"00a6ecd7-a808-43a6-9562-953818cc9020\") " Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.893996 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00a6ecd7-a808-43a6-9562-953818cc9020-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00a6ecd7-a808-43a6-9562-953818cc9020" (UID: "00a6ecd7-a808-43a6-9562-953818cc9020"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.919321 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a6ecd7-a808-43a6-9562-953818cc9020-kube-api-access-8mqvl" (OuterVolumeSpecName: "kube-api-access-8mqvl") pod "00a6ecd7-a808-43a6-9562-953818cc9020" (UID: "00a6ecd7-a808-43a6-9562-953818cc9020"). InnerVolumeSpecName "kube-api-access-8mqvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.935895 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8bd6e13-5a1a-431a-b28a-08aee6a5f619" path="/var/lib/kubelet/pods/b8bd6e13-5a1a-431a-b28a-08aee6a5f619/volumes" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.995466 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mqvl\" (UniqueName: \"kubernetes.io/projected/00a6ecd7-a808-43a6-9562-953818cc9020-kube-api-access-8mqvl\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:40 crc kubenswrapper[4679]: I1202 10:37:40.995502 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a6ecd7-a808-43a6-9562-953818cc9020-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.134249 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.150685 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.157969 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.171199 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.185822 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299709 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5p4g\" (UniqueName: \"kubernetes.io/projected/db8eb5a7-5fef-4f4a-905f-40103b0f513d-kube-api-access-n5p4g\") pod \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299779 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pbk5\" (UniqueName: \"kubernetes.io/projected/59eae751-1d2f-4ce3-a386-d2dcc0727717-kube-api-access-8pbk5\") pod \"59eae751-1d2f-4ce3-a386-d2dcc0727717\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299811 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjzkr\" (UniqueName: \"kubernetes.io/projected/c2f792c0-0b54-4599-aea3-d5567072e6f7-kube-api-access-hjzkr\") pod \"c2f792c0-0b54-4599-aea3-d5567072e6f7\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299852 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59eae751-1d2f-4ce3-a386-d2dcc0727717-operator-scripts\") pod \"59eae751-1d2f-4ce3-a386-d2dcc0727717\" (UID: \"59eae751-1d2f-4ce3-a386-d2dcc0727717\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299886 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6f8df03-1845-4c2b-a30a-3d8186287f73-operator-scripts\") pod \"e6f8df03-1845-4c2b-a30a-3d8186287f73\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299922 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-operator-scripts\") pod \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299964 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcngt\" (UniqueName: \"kubernetes.io/projected/e6f8df03-1845-4c2b-a30a-3d8186287f73-kube-api-access-mcngt\") pod \"e6f8df03-1845-4c2b-a30a-3d8186287f73\" (UID: \"e6f8df03-1845-4c2b-a30a-3d8186287f73\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.299990 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2f792c0-0b54-4599-aea3-d5567072e6f7-operator-scripts\") pod \"c2f792c0-0b54-4599-aea3-d5567072e6f7\" (UID: \"c2f792c0-0b54-4599-aea3-d5567072e6f7\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.300036 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db8eb5a7-5fef-4f4a-905f-40103b0f513d-operator-scripts\") pod \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\" (UID: \"db8eb5a7-5fef-4f4a-905f-40103b0f513d\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.300077 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fffms\" (UniqueName: \"kubernetes.io/projected/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-kube-api-access-fffms\") pod \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\" (UID: \"f5c4d5af-bdac-4486-b0ab-f389cd3331bb\") " Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.301259 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59eae751-1d2f-4ce3-a386-d2dcc0727717-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "59eae751-1d2f-4ce3-a386-d2dcc0727717" (UID: "59eae751-1d2f-4ce3-a386-d2dcc0727717"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.301444 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f792c0-0b54-4599-aea3-d5567072e6f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2f792c0-0b54-4599-aea3-d5567072e6f7" (UID: "c2f792c0-0b54-4599-aea3-d5567072e6f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.301489 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db8eb5a7-5fef-4f4a-905f-40103b0f513d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db8eb5a7-5fef-4f4a-905f-40103b0f513d" (UID: "db8eb5a7-5fef-4f4a-905f-40103b0f513d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.301677 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5c4d5af-bdac-4486-b0ab-f389cd3331bb" (UID: "f5c4d5af-bdac-4486-b0ab-f389cd3331bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.302120 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6f8df03-1845-4c2b-a30a-3d8186287f73-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6f8df03-1845-4c2b-a30a-3d8186287f73" (UID: "e6f8df03-1845-4c2b-a30a-3d8186287f73"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.305746 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db8eb5a7-5fef-4f4a-905f-40103b0f513d-kube-api-access-n5p4g" (OuterVolumeSpecName: "kube-api-access-n5p4g") pod "db8eb5a7-5fef-4f4a-905f-40103b0f513d" (UID: "db8eb5a7-5fef-4f4a-905f-40103b0f513d"). InnerVolumeSpecName "kube-api-access-n5p4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.305904 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6f8df03-1845-4c2b-a30a-3d8186287f73-kube-api-access-mcngt" (OuterVolumeSpecName: "kube-api-access-mcngt") pod "e6f8df03-1845-4c2b-a30a-3d8186287f73" (UID: "e6f8df03-1845-4c2b-a30a-3d8186287f73"). InnerVolumeSpecName "kube-api-access-mcngt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.305939 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f792c0-0b54-4599-aea3-d5567072e6f7-kube-api-access-hjzkr" (OuterVolumeSpecName: "kube-api-access-hjzkr") pod "c2f792c0-0b54-4599-aea3-d5567072e6f7" (UID: "c2f792c0-0b54-4599-aea3-d5567072e6f7"). InnerVolumeSpecName "kube-api-access-hjzkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.305956 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-kube-api-access-fffms" (OuterVolumeSpecName: "kube-api-access-fffms") pod "f5c4d5af-bdac-4486-b0ab-f389cd3331bb" (UID: "f5c4d5af-bdac-4486-b0ab-f389cd3331bb"). InnerVolumeSpecName "kube-api-access-fffms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.306154 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59eae751-1d2f-4ce3-a386-d2dcc0727717-kube-api-access-8pbk5" (OuterVolumeSpecName: "kube-api-access-8pbk5") pod "59eae751-1d2f-4ce3-a386-d2dcc0727717" (UID: "59eae751-1d2f-4ce3-a386-d2dcc0727717"). InnerVolumeSpecName "kube-api-access-8pbk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402283 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db8eb5a7-5fef-4f4a-905f-40103b0f513d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402403 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fffms\" (UniqueName: \"kubernetes.io/projected/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-kube-api-access-fffms\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402418 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5p4g\" (UniqueName: \"kubernetes.io/projected/db8eb5a7-5fef-4f4a-905f-40103b0f513d-kube-api-access-n5p4g\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402430 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pbk5\" (UniqueName: \"kubernetes.io/projected/59eae751-1d2f-4ce3-a386-d2dcc0727717-kube-api-access-8pbk5\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402443 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjzkr\" (UniqueName: \"kubernetes.io/projected/c2f792c0-0b54-4599-aea3-d5567072e6f7-kube-api-access-hjzkr\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402453 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59eae751-1d2f-4ce3-a386-d2dcc0727717-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402464 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6f8df03-1845-4c2b-a30a-3d8186287f73-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402474 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c4d5af-bdac-4486-b0ab-f389cd3331bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402484 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcngt\" (UniqueName: \"kubernetes.io/projected/e6f8df03-1845-4c2b-a30a-3d8186287f73-kube-api-access-mcngt\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.402494 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2f792c0-0b54-4599-aea3-d5567072e6f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.430441 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bf39-account-create-update-9gvms" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.430429 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bf39-account-create-update-9gvms" event={"ID":"f5c4d5af-bdac-4486-b0ab-f389cd3331bb","Type":"ContainerDied","Data":"044181bd3fb93163017fc824b261ec1f0e29f1bb66ce32c11ef1cb82015fc7f2"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.430508 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="044181bd3fb93163017fc824b261ec1f0e29f1bb66ce32c11ef1cb82015fc7f2" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.433107 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"416d04f1-7afb-4bdf-a5d5-1a0c381513f9","Type":"ContainerStarted","Data":"0258d496799a824dccc899444561afc95538f8181fcb3bbbe5c69aa6d0d1d35c"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.433377 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.435182 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2xs5t" event={"ID":"db8eb5a7-5fef-4f4a-905f-40103b0f513d","Type":"ContainerDied","Data":"c0634c68c8e98f2d82e66fd30fafb166ea33a8e2e88b336072cdbd6bb4a43fa7"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.435215 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0634c68c8e98f2d82e66fd30fafb166ea33a8e2e88b336072cdbd6bb4a43fa7" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.435257 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2xs5t" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.440510 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w84l8" event={"ID":"00a6ecd7-a808-43a6-9562-953818cc9020","Type":"ContainerDied","Data":"9ac1ebd282cb365ee3103b8ddc424b00ef36597dabdd42e2395a349bf5a19535"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.440637 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ac1ebd282cb365ee3103b8ddc424b00ef36597dabdd42e2395a349bf5a19535" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.440528 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w84l8" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.442354 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-164a-account-create-update-bsvts" event={"ID":"59eae751-1d2f-4ce3-a386-d2dcc0727717","Type":"ContainerDied","Data":"668ff710ea936e4f69bdde67d4a98b218d8c82afbc5ea09badef2847b75bad95"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.442390 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="668ff710ea936e4f69bdde67d4a98b218d8c82afbc5ea09badef2847b75bad95" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.442440 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-164a-account-create-update-bsvts" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.448631 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e4cf-account-create-update-fp8rh" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.449436 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e4cf-account-create-update-fp8rh" event={"ID":"c2f792c0-0b54-4599-aea3-d5567072e6f7","Type":"ContainerDied","Data":"d79bd1cc45276bbbe513c494d077f08326eb92a502d0b55c4a8d47fb468c10f4"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.449485 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d79bd1cc45276bbbe513c494d077f08326eb92a502d0b55c4a8d47fb468c10f4" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.451675 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w6dwk" event={"ID":"e6f8df03-1845-4c2b-a30a-3d8186287f73","Type":"ContainerDied","Data":"41a255764934c57806da9b3b4577c3489a8aa91dc6700a5b7934f3216fc206d2"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.451717 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41a255764934c57806da9b3b4577c3489a8aa91dc6700a5b7934f3216fc206d2" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.452107 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w6dwk" Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.454478 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"a3c9d8b821b44d3aa1e0b3f6a713b46b0ced7f5da35cd114fd0dc15aaa3484a1"} Dec 02 10:37:41 crc kubenswrapper[4679]: I1202 10:37:41.479830 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371959.374966 podStartE2EDuration="1m17.479810605s" podCreationTimestamp="2025-12-02 10:36:24 +0000 UTC" firstStartedPulling="2025-12-02 10:36:26.595602762 +0000 UTC m=+1039.925741632" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:37:41.469043275 +0000 UTC m=+1114.799182135" watchObservedRunningTime="2025-12-02 10:37:41.479810605 +0000 UTC m=+1114.809949465" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.525815 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-zs6tz"] Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526736 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59eae751-1d2f-4ce3-a386-d2dcc0727717" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526755 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="59eae751-1d2f-4ce3-a386-d2dcc0727717" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526777 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2f792c0-0b54-4599-aea3-d5567072e6f7" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526784 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2f792c0-0b54-4599-aea3-d5567072e6f7" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526798 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db8eb5a7-5fef-4f4a-905f-40103b0f513d" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526806 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="db8eb5a7-5fef-4f4a-905f-40103b0f513d" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526822 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f8df03-1845-4c2b-a30a-3d8186287f73" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526829 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f8df03-1845-4c2b-a30a-3d8186287f73" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526844 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bd6e13-5a1a-431a-b28a-08aee6a5f619" containerName="ovn-config" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526850 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bd6e13-5a1a-431a-b28a-08aee6a5f619" containerName="ovn-config" Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526864 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5c4d5af-bdac-4486-b0ab-f389cd3331bb" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526871 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5c4d5af-bdac-4486-b0ab-f389cd3331bb" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: E1202 10:37:44.526883 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a6ecd7-a808-43a6-9562-953818cc9020" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.526890 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a6ecd7-a808-43a6-9562-953818cc9020" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527087 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5c4d5af-bdac-4486-b0ab-f389cd3331bb" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527125 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="db8eb5a7-5fef-4f4a-905f-40103b0f513d" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527142 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2f792c0-0b54-4599-aea3-d5567072e6f7" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527165 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="59eae751-1d2f-4ce3-a386-d2dcc0727717" containerName="mariadb-account-create-update" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527187 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8bd6e13-5a1a-431a-b28a-08aee6a5f619" containerName="ovn-config" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527198 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6f8df03-1845-4c2b-a30a-3d8186287f73" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527213 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a6ecd7-a808-43a6-9562-953818cc9020" containerName="mariadb-database-create" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.527848 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.530151 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.530320 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-48wtd" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.538910 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zs6tz"] Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.678557 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-combined-ca-bundle\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.678610 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m69r2\" (UniqueName: \"kubernetes.io/projected/8d59898e-672f-4699-bf31-a964cd84913b-kube-api-access-m69r2\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.678643 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-db-sync-config-data\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.678888 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-config-data\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.780504 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-config-data\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.780586 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-combined-ca-bundle\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.780606 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m69r2\" (UniqueName: \"kubernetes.io/projected/8d59898e-672f-4699-bf31-a964cd84913b-kube-api-access-m69r2\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.780630 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-db-sync-config-data\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.787396 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-db-sync-config-data\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.787620 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-combined-ca-bundle\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.789974 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-config-data\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.801082 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m69r2\" (UniqueName: \"kubernetes.io/projected/8d59898e-672f-4699-bf31-a964cd84913b-kube-api-access-m69r2\") pod \"glance-db-sync-zs6tz\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:44 crc kubenswrapper[4679]: I1202 10:37:44.852009 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zs6tz" Dec 02 10:37:45 crc kubenswrapper[4679]: I1202 10:37:45.689213 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zs6tz"] Dec 02 10:37:46 crc kubenswrapper[4679]: I1202 10:37:46.511084 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"65b11751fcac0e196e8134fcab36d92d9cb825595d66c2d7b1c5b69ae1af9d8f"} Dec 02 10:37:46 crc kubenswrapper[4679]: I1202 10:37:46.511159 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"78ec80b77fa85cd50353e1f00258ddc0f691a679c8d6fcb1b39bbfa49436b40d"} Dec 02 10:37:46 crc kubenswrapper[4679]: I1202 10:37:46.511175 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"e933cf92d9915bef17f993a7483f74e3ceeae27138f09fa1df5b61d741a5874f"} Dec 02 10:37:46 crc kubenswrapper[4679]: I1202 10:37:46.512516 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zs6tz" event={"ID":"8d59898e-672f-4699-bf31-a964cd84913b","Type":"ContainerStarted","Data":"79e47463117cf88a4f39a1c42956583d14fb2c06d9aa46a172fffb2481537a0c"} Dec 02 10:37:47 crc kubenswrapper[4679]: I1202 10:37:47.527709 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"b5d40f987cce500818d335902150b3f69fa44b4e6cc9844ddb72799ea3cb241a"} Dec 02 10:37:50 crc kubenswrapper[4679]: I1202 10:37:50.556294 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"aeb2bc1da6707aa4b63e2a3c5faf74d9df6dcd0f1e70797a68d80b14d4445ded"} Dec 02 10:37:50 crc kubenswrapper[4679]: I1202 10:37:50.557910 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"b2ccd78852134e6297f3d3ac0fe7ae13f2f1fcf46e9b78ad33f4576742bb8b33"} Dec 02 10:37:50 crc kubenswrapper[4679]: I1202 10:37:50.557993 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"4f452b989ed8698b7673b2ffaaaed1882ccce0de402549e4aa264f053f3379f6"} Dec 02 10:37:50 crc kubenswrapper[4679]: I1202 10:37:50.558075 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"1d5966e1c71ba49170b31767aa4ddcb51d398bbb53a75e7c4931c52e76dd4856"} Dec 02 10:37:51 crc kubenswrapper[4679]: I1202 10:37:51.570357 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"0219df43739381dd2b28e36c41db83a0bb36ee75e0333dd9fca4ee844a4a63ce"} Dec 02 10:37:51 crc kubenswrapper[4679]: I1202 10:37:51.570405 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"9f1411dd5c7ff07babc7558995acd740c069c8dc42784399044096c89195717a"} Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.512494 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.767521 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-ctlv8"] Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.768507 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.780990 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ctlv8"] Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.854429 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dflbc\" (UniqueName: \"kubernetes.io/projected/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-kube-api-access-dflbc\") pod \"cinder-db-create-ctlv8\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.854488 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-operator-scripts\") pod \"cinder-db-create-ctlv8\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.884329 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qbg4x"] Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.885506 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.901242 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b3a9-account-create-update-jg27q"] Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.902224 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.904545 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.924518 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qbg4x"] Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.950514 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b3a9-account-create-update-jg27q"] Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.965131 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0969458b-57a6-4ec6-92b9-870ee1cfc6da-operator-scripts\") pod \"barbican-db-create-qbg4x\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.965219 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dflbc\" (UniqueName: \"kubernetes.io/projected/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-kube-api-access-dflbc\") pod \"cinder-db-create-ctlv8\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.965245 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-operator-scripts\") pod \"cinder-db-create-ctlv8\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.965271 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvlr9\" (UniqueName: \"kubernetes.io/projected/0969458b-57a6-4ec6-92b9-870ee1cfc6da-kube-api-access-lvlr9\") pod \"barbican-db-create-qbg4x\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.966255 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-operator-scripts\") pod \"cinder-db-create-ctlv8\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:55 crc kubenswrapper[4679]: I1202 10:37:55.993416 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.024156 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-46e9-account-create-update-dbkf8"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.025220 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.036170 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.036490 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-46e9-account-create-update-dbkf8"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.039143 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dflbc\" (UniqueName: \"kubernetes.io/projected/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-kube-api-access-dflbc\") pod \"cinder-db-create-ctlv8\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.067231 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0969458b-57a6-4ec6-92b9-870ee1cfc6da-operator-scripts\") pod \"barbican-db-create-qbg4x\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.067653 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/595fb7b3-d70c-4cf2-98c4-287fff3ec997-operator-scripts\") pod \"cinder-b3a9-account-create-update-jg27q\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.067724 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvlr9\" (UniqueName: \"kubernetes.io/projected/0969458b-57a6-4ec6-92b9-870ee1cfc6da-kube-api-access-lvlr9\") pod \"barbican-db-create-qbg4x\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.067779 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng4bf\" (UniqueName: \"kubernetes.io/projected/595fb7b3-d70c-4cf2-98c4-287fff3ec997-kube-api-access-ng4bf\") pod \"cinder-b3a9-account-create-update-jg27q\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.070454 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0969458b-57a6-4ec6-92b9-870ee1cfc6da-operator-scripts\") pod \"barbican-db-create-qbg4x\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.098582 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ctlv8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.099914 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvlr9\" (UniqueName: \"kubernetes.io/projected/0969458b-57a6-4ec6-92b9-870ee1cfc6da-kube-api-access-lvlr9\") pod \"barbican-db-create-qbg4x\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.169497 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/595fb7b3-d70c-4cf2-98c4-287fff3ec997-operator-scripts\") pod \"cinder-b3a9-account-create-update-jg27q\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.169614 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng4bf\" (UniqueName: \"kubernetes.io/projected/595fb7b3-d70c-4cf2-98c4-287fff3ec997-kube-api-access-ng4bf\") pod \"cinder-b3a9-account-create-update-jg27q\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.169696 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95mdf\" (UniqueName: \"kubernetes.io/projected/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-kube-api-access-95mdf\") pod \"barbican-46e9-account-create-update-dbkf8\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.169779 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-operator-scripts\") pod \"barbican-46e9-account-create-update-dbkf8\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.173143 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/595fb7b3-d70c-4cf2-98c4-287fff3ec997-operator-scripts\") pod \"cinder-b3a9-account-create-update-jg27q\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.183172 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-dk9q8"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.184401 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.187633 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.187830 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5x6pg" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.187976 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.188114 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.194499 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dk9q8"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.205257 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qbg4x" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.217549 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-mjmmw"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.225339 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng4bf\" (UniqueName: \"kubernetes.io/projected/595fb7b3-d70c-4cf2-98c4-287fff3ec997-kube-api-access-ng4bf\") pod \"cinder-b3a9-account-create-update-jg27q\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.226348 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.240462 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mjmmw"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.300647 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-operator-scripts\") pod \"barbican-46e9-account-create-update-dbkf8\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.300748 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d57d916-49db-43e8-b1cd-7bd5406d4812-operator-scripts\") pod \"neutron-db-create-mjmmw\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.300812 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-combined-ca-bundle\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.300859 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krtgm\" (UniqueName: \"kubernetes.io/projected/9f9bcb09-6b18-4357-b474-0db2b33e1641-kube-api-access-krtgm\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.301072 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtwcn\" (UniqueName: \"kubernetes.io/projected/0d57d916-49db-43e8-b1cd-7bd5406d4812-kube-api-access-gtwcn\") pod \"neutron-db-create-mjmmw\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.301089 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-operator-scripts\") pod \"barbican-46e9-account-create-update-dbkf8\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.302006 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-config-data\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.302060 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95mdf\" (UniqueName: \"kubernetes.io/projected/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-kube-api-access-95mdf\") pod \"barbican-46e9-account-create-update-dbkf8\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.322754 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e045-account-create-update-c77tp"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.323758 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.330667 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.335315 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e045-account-create-update-c77tp"] Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.360015 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95mdf\" (UniqueName: \"kubernetes.io/projected/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-kube-api-access-95mdf\") pod \"barbican-46e9-account-create-update-dbkf8\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.378681 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404275 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-combined-ca-bundle\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404354 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krtgm\" (UniqueName: \"kubernetes.io/projected/9f9bcb09-6b18-4357-b474-0db2b33e1641-kube-api-access-krtgm\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404443 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtwcn\" (UniqueName: \"kubernetes.io/projected/0d57d916-49db-43e8-b1cd-7bd5406d4812-kube-api-access-gtwcn\") pod \"neutron-db-create-mjmmw\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404506 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-config-data\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404583 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5b4081a-2347-45b3-b5db-07de3d5a90c2-operator-scripts\") pod \"neutron-e045-account-create-update-c77tp\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404619 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d57d916-49db-43e8-b1cd-7bd5406d4812-operator-scripts\") pod \"neutron-db-create-mjmmw\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.404642 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6wzq\" (UniqueName: \"kubernetes.io/projected/c5b4081a-2347-45b3-b5db-07de3d5a90c2-kube-api-access-x6wzq\") pod \"neutron-e045-account-create-update-c77tp\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.408008 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-config-data\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.408174 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d57d916-49db-43e8-b1cd-7bd5406d4812-operator-scripts\") pod \"neutron-db-create-mjmmw\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.408941 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-combined-ca-bundle\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.424051 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krtgm\" (UniqueName: \"kubernetes.io/projected/9f9bcb09-6b18-4357-b474-0db2b33e1641-kube-api-access-krtgm\") pod \"keystone-db-sync-dk9q8\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.424521 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtwcn\" (UniqueName: \"kubernetes.io/projected/0d57d916-49db-43e8-b1cd-7bd5406d4812-kube-api-access-gtwcn\") pod \"neutron-db-create-mjmmw\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.506846 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5b4081a-2347-45b3-b5db-07de3d5a90c2-operator-scripts\") pod \"neutron-e045-account-create-update-c77tp\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.506907 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6wzq\" (UniqueName: \"kubernetes.io/projected/c5b4081a-2347-45b3-b5db-07de3d5a90c2-kube-api-access-x6wzq\") pod \"neutron-e045-account-create-update-c77tp\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.509532 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5b4081a-2347-45b3-b5db-07de3d5a90c2-operator-scripts\") pod \"neutron-e045-account-create-update-c77tp\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.518633 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.525853 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6wzq\" (UniqueName: \"kubernetes.io/projected/c5b4081a-2347-45b3-b5db-07de3d5a90c2-kube-api-access-x6wzq\") pod \"neutron-e045-account-create-update-c77tp\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.548501 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.602670 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mjmmw" Dec 02 10:37:56 crc kubenswrapper[4679]: I1202 10:37:56.683665 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:38:04 crc kubenswrapper[4679]: E1202 10:38:04.116668 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 02 10:38:04 crc kubenswrapper[4679]: E1202 10:38:04.117501 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m69r2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-zs6tz_openstack(8d59898e-672f-4699-bf31-a964cd84913b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:04 crc kubenswrapper[4679]: E1202 10:38:04.118876 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-zs6tz" podUID="8d59898e-672f-4699-bf31-a964cd84913b" Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.694078 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d8428a26-75d9-4d87-b31d-79329264e14e","Type":"ContainerStarted","Data":"bd26bd3d0cc0d7d7b904df8b27245d97db0651d4b722e9f9a87719e5c3637b19"} Dec 02 10:38:04 crc kubenswrapper[4679]: E1202 10:38:04.694601 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-zs6tz" podUID="8d59898e-672f-4699-bf31-a964cd84913b" Dec 02 10:38:04 crc kubenswrapper[4679]: W1202 10:38:04.699512 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb4c80f2_bd41_4e94_b8e9_c8d6616357e6.slice/crio-039f0079010093bdcb3ab0b055ef144ad6edc7c23e1f94e75dfb63cf0c75f46f WatchSource:0}: Error finding container 039f0079010093bdcb3ab0b055ef144ad6edc7c23e1f94e75dfb63cf0c75f46f: Status 404 returned error can't find the container with id 039f0079010093bdcb3ab0b055ef144ad6edc7c23e1f94e75dfb63cf0c75f46f Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.719430 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mjmmw"] Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.741596 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-46e9-account-create-update-dbkf8"] Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.771726 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dk9q8"] Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.779806 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=49.105957362 podStartE2EDuration="1m3.779780209s" podCreationTimestamp="2025-12-02 10:37:01 +0000 UTC" firstStartedPulling="2025-12-02 10:37:34.913632101 +0000 UTC m=+1108.243770961" lastFinishedPulling="2025-12-02 10:37:49.587454948 +0000 UTC m=+1122.917593808" observedRunningTime="2025-12-02 10:38:04.750237569 +0000 UTC m=+1138.080376439" watchObservedRunningTime="2025-12-02 10:38:04.779780209 +0000 UTC m=+1138.109919079" Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.844112 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b3a9-account-create-update-jg27q"] Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.850809 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ctlv8"] Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.857201 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e045-account-create-update-c77tp"] Dec 02 10:38:04 crc kubenswrapper[4679]: I1202 10:38:04.864056 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qbg4x"] Dec 02 10:38:04 crc kubenswrapper[4679]: W1202 10:38:04.865872 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5b4081a_2347_45b3_b5db_07de3d5a90c2.slice/crio-0aad27b66df05b49626774a793934d00186dc98c5c7f10494eaca54ca206e645 WatchSource:0}: Error finding container 0aad27b66df05b49626774a793934d00186dc98c5c7f10494eaca54ca206e645: Status 404 returned error can't find the container with id 0aad27b66df05b49626774a793934d00186dc98c5c7f10494eaca54ca206e645 Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.019320 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sv4sp"] Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.020953 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.024804 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.033573 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sv4sp"] Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.079386 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-config\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.079430 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.079506 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwssg\" (UniqueName: \"kubernetes.io/projected/11db932d-1c7f-4bdf-8400-47c562d2cff7-kube-api-access-fwssg\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.079523 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.079586 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.079624 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.180934 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.180990 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-config\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.181006 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.181075 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwssg\" (UniqueName: \"kubernetes.io/projected/11db932d-1c7f-4bdf-8400-47c562d2cff7-kube-api-access-fwssg\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.181093 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.181140 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.182637 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.182961 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-config\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.183005 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.183125 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.183451 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.203094 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwssg\" (UniqueName: \"kubernetes.io/projected/11db932d-1c7f-4bdf-8400-47c562d2cff7-kube-api-access-fwssg\") pod \"dnsmasq-dns-77585f5f8c-sv4sp\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.359870 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.703450 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b3a9-account-create-update-jg27q" event={"ID":"595fb7b3-d70c-4cf2-98c4-287fff3ec997","Type":"ContainerStarted","Data":"59132a40f0f7bf6fb36a5e0df1bd62d1c7827c4155601e5c562af3e1e31536ec"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.703859 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b3a9-account-create-update-jg27q" event={"ID":"595fb7b3-d70c-4cf2-98c4-287fff3ec997","Type":"ContainerStarted","Data":"7bf7ac41a8d5ba915f76e2851b76d8525cddab58be9db462638f15e16117c5f8"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.706031 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mjmmw" event={"ID":"0d57d916-49db-43e8-b1cd-7bd5406d4812","Type":"ContainerStarted","Data":"3a6a97813d3d5e88564b30008c92dc07ec16c17d3e6c1a407fe5df9d448f2c9d"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.706105 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mjmmw" event={"ID":"0d57d916-49db-43e8-b1cd-7bd5406d4812","Type":"ContainerStarted","Data":"854c4f13033338080c90c19fc4fa166ebedda934b647ddc48f77ddcdb4f14a17"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.707435 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dk9q8" event={"ID":"9f9bcb09-6b18-4357-b474-0db2b33e1641","Type":"ContainerStarted","Data":"3b406f25b4ee23e589aab395dfc2adf230fd7dbde0bf8d20cf5e35b360d6ff7d"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.713065 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ctlv8" event={"ID":"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040","Type":"ContainerStarted","Data":"b7d2a2ad8f04c9cd1576165896296ca93185f6babc3d040bc804ca53afd1d7b3"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.713103 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ctlv8" event={"ID":"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040","Type":"ContainerStarted","Data":"e969d76211bbcf8a1bb287f5680b90f666fe4d1fec14238eac2f0bffb77d9a83"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.714811 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e045-account-create-update-c77tp" event={"ID":"c5b4081a-2347-45b3-b5db-07de3d5a90c2","Type":"ContainerStarted","Data":"424e1d8b153bdce6f42ed6328cdaa50f7c461b3e8bccc32a8b19058b7892596c"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.714858 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e045-account-create-update-c77tp" event={"ID":"c5b4081a-2347-45b3-b5db-07de3d5a90c2","Type":"ContainerStarted","Data":"0aad27b66df05b49626774a793934d00186dc98c5c7f10494eaca54ca206e645"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.716244 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qbg4x" event={"ID":"0969458b-57a6-4ec6-92b9-870ee1cfc6da","Type":"ContainerStarted","Data":"ae341e015d059b83e96e1f4f910fcc9ad91d74d843c59ab7a7035273a6fdd1cc"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.716285 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qbg4x" event={"ID":"0969458b-57a6-4ec6-92b9-870ee1cfc6da","Type":"ContainerStarted","Data":"c9dd3f4fce58c702db25407343befe5f0385c5ddad4499076a3d1b6b16807812"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.721425 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-b3a9-account-create-update-jg27q" podStartSLOduration=10.721410412000001 podStartE2EDuration="10.721410412s" podCreationTimestamp="2025-12-02 10:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:05.720878017 +0000 UTC m=+1139.051016877" watchObservedRunningTime="2025-12-02 10:38:05.721410412 +0000 UTC m=+1139.051549262" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.731430 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-46e9-account-create-update-dbkf8" event={"ID":"db4c80f2-bd41-4e94-b8e9-c8d6616357e6","Type":"ContainerStarted","Data":"b8610e7c12a0b7efd7c39c1944ed1b04aadc67ee634cf5d4dae991e3870e82f4"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.731487 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-46e9-account-create-update-dbkf8" event={"ID":"db4c80f2-bd41-4e94-b8e9-c8d6616357e6","Type":"ContainerStarted","Data":"039f0079010093bdcb3ab0b055ef144ad6edc7c23e1f94e75dfb63cf0c75f46f"} Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.739517 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-e045-account-create-update-c77tp" podStartSLOduration=9.739498923 podStartE2EDuration="9.739498923s" podCreationTimestamp="2025-12-02 10:37:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:05.737884406 +0000 UTC m=+1139.068023276" watchObservedRunningTime="2025-12-02 10:38:05.739498923 +0000 UTC m=+1139.069637783" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.772056 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-ctlv8" podStartSLOduration=10.772039279 podStartE2EDuration="10.772039279s" podCreationTimestamp="2025-12-02 10:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:05.762613308 +0000 UTC m=+1139.092752168" watchObservedRunningTime="2025-12-02 10:38:05.772039279 +0000 UTC m=+1139.102178139" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.788791 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-qbg4x" podStartSLOduration=10.788771761 podStartE2EDuration="10.788771761s" podCreationTimestamp="2025-12-02 10:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:05.777836926 +0000 UTC m=+1139.107975786" watchObservedRunningTime="2025-12-02 10:38:05.788771761 +0000 UTC m=+1139.118910631" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.801783 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-46e9-account-create-update-dbkf8" podStartSLOduration=10.801765065 podStartE2EDuration="10.801765065s" podCreationTimestamp="2025-12-02 10:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:05.792452167 +0000 UTC m=+1139.122591027" watchObservedRunningTime="2025-12-02 10:38:05.801765065 +0000 UTC m=+1139.131903925" Dec 02 10:38:05 crc kubenswrapper[4679]: I1202 10:38:05.878580 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sv4sp"] Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.742296 4679 generic.go:334] "Generic (PLEG): container finished" podID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerID="76c7d26940b8801e3d55aa1126f9cb0e60ef7aa2d32d3cfd779ee958041e86b9" exitCode=0 Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.742507 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" event={"ID":"11db932d-1c7f-4bdf-8400-47c562d2cff7","Type":"ContainerDied","Data":"76c7d26940b8801e3d55aa1126f9cb0e60ef7aa2d32d3cfd779ee958041e86b9"} Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.742757 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" event={"ID":"11db932d-1c7f-4bdf-8400-47c562d2cff7","Type":"ContainerStarted","Data":"b5d1940693da8a5a4204af0097908e9e02284e6c177c57a5161dad2c4cc59993"} Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.764279 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ctlv8" event={"ID":"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040","Type":"ContainerDied","Data":"b7d2a2ad8f04c9cd1576165896296ca93185f6babc3d040bc804ca53afd1d7b3"} Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.764248 4679 generic.go:334] "Generic (PLEG): container finished" podID="f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" containerID="b7d2a2ad8f04c9cd1576165896296ca93185f6babc3d040bc804ca53afd1d7b3" exitCode=0 Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.776403 4679 generic.go:334] "Generic (PLEG): container finished" podID="0969458b-57a6-4ec6-92b9-870ee1cfc6da" containerID="ae341e015d059b83e96e1f4f910fcc9ad91d74d843c59ab7a7035273a6fdd1cc" exitCode=0 Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.776541 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qbg4x" event={"ID":"0969458b-57a6-4ec6-92b9-870ee1cfc6da","Type":"ContainerDied","Data":"ae341e015d059b83e96e1f4f910fcc9ad91d74d843c59ab7a7035273a6fdd1cc"} Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.780556 4679 generic.go:334] "Generic (PLEG): container finished" podID="db4c80f2-bd41-4e94-b8e9-c8d6616357e6" containerID="b8610e7c12a0b7efd7c39c1944ed1b04aadc67ee634cf5d4dae991e3870e82f4" exitCode=0 Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.780637 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-46e9-account-create-update-dbkf8" event={"ID":"db4c80f2-bd41-4e94-b8e9-c8d6616357e6","Type":"ContainerDied","Data":"b8610e7c12a0b7efd7c39c1944ed1b04aadc67ee634cf5d4dae991e3870e82f4"} Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.784848 4679 generic.go:334] "Generic (PLEG): container finished" podID="595fb7b3-d70c-4cf2-98c4-287fff3ec997" containerID="59132a40f0f7bf6fb36a5e0df1bd62d1c7827c4155601e5c562af3e1e31536ec" exitCode=0 Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.784899 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b3a9-account-create-update-jg27q" event={"ID":"595fb7b3-d70c-4cf2-98c4-287fff3ec997","Type":"ContainerDied","Data":"59132a40f0f7bf6fb36a5e0df1bd62d1c7827c4155601e5c562af3e1e31536ec"} Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.788265 4679 generic.go:334] "Generic (PLEG): container finished" podID="0d57d916-49db-43e8-b1cd-7bd5406d4812" containerID="3a6a97813d3d5e88564b30008c92dc07ec16c17d3e6c1a407fe5df9d448f2c9d" exitCode=0 Dec 02 10:38:06 crc kubenswrapper[4679]: I1202 10:38:06.788491 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mjmmw" event={"ID":"0d57d916-49db-43e8-b1cd-7bd5406d4812","Type":"ContainerDied","Data":"3a6a97813d3d5e88564b30008c92dc07ec16c17d3e6c1a407fe5df9d448f2c9d"} Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.153287 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mjmmw" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.367856 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtwcn\" (UniqueName: \"kubernetes.io/projected/0d57d916-49db-43e8-b1cd-7bd5406d4812-kube-api-access-gtwcn\") pod \"0d57d916-49db-43e8-b1cd-7bd5406d4812\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.367926 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d57d916-49db-43e8-b1cd-7bd5406d4812-operator-scripts\") pod \"0d57d916-49db-43e8-b1cd-7bd5406d4812\" (UID: \"0d57d916-49db-43e8-b1cd-7bd5406d4812\") " Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.368396 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d57d916-49db-43e8-b1cd-7bd5406d4812-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d57d916-49db-43e8-b1cd-7bd5406d4812" (UID: "0d57d916-49db-43e8-b1cd-7bd5406d4812"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.372941 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d57d916-49db-43e8-b1cd-7bd5406d4812-kube-api-access-gtwcn" (OuterVolumeSpecName: "kube-api-access-gtwcn") pod "0d57d916-49db-43e8-b1cd-7bd5406d4812" (UID: "0d57d916-49db-43e8-b1cd-7bd5406d4812"). InnerVolumeSpecName "kube-api-access-gtwcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.469466 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d57d916-49db-43e8-b1cd-7bd5406d4812-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.469499 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtwcn\" (UniqueName: \"kubernetes.io/projected/0d57d916-49db-43e8-b1cd-7bd5406d4812-kube-api-access-gtwcn\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.799186 4679 generic.go:334] "Generic (PLEG): container finished" podID="c5b4081a-2347-45b3-b5db-07de3d5a90c2" containerID="424e1d8b153bdce6f42ed6328cdaa50f7c461b3e8bccc32a8b19058b7892596c" exitCode=0 Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.799300 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e045-account-create-update-c77tp" event={"ID":"c5b4081a-2347-45b3-b5db-07de3d5a90c2","Type":"ContainerDied","Data":"424e1d8b153bdce6f42ed6328cdaa50f7c461b3e8bccc32a8b19058b7892596c"} Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.802053 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mjmmw" event={"ID":"0d57d916-49db-43e8-b1cd-7bd5406d4812","Type":"ContainerDied","Data":"854c4f13033338080c90c19fc4fa166ebedda934b647ddc48f77ddcdb4f14a17"} Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.802091 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="854c4f13033338080c90c19fc4fa166ebedda934b647ddc48f77ddcdb4f14a17" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.802140 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mjmmw" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.805822 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" event={"ID":"11db932d-1c7f-4bdf-8400-47c562d2cff7","Type":"ContainerStarted","Data":"c53e9b861fa7f5dd3cf4c7650b06bf990c2be32822c8b90dfbc5af689f11ba1d"} Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.805872 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:07 crc kubenswrapper[4679]: I1202 10:38:07.839713 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podStartSLOduration=2.839691242 podStartE2EDuration="2.839691242s" podCreationTimestamp="2025-12-02 10:38:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:07.838684163 +0000 UTC m=+1141.168823033" watchObservedRunningTime="2025-12-02 10:38:07.839691242 +0000 UTC m=+1141.169830112" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.089512 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qbg4x" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.097386 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ctlv8" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.110167 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.136700 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.137858 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.218349 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dflbc\" (UniqueName: \"kubernetes.io/projected/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-kube-api-access-dflbc\") pod \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.218552 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95mdf\" (UniqueName: \"kubernetes.io/projected/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-kube-api-access-95mdf\") pod \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.218583 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0969458b-57a6-4ec6-92b9-870ee1cfc6da-operator-scripts\") pod \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.218622 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-operator-scripts\") pod \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\" (UID: \"db4c80f2-bd41-4e94-b8e9-c8d6616357e6\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.218662 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-operator-scripts\") pod \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\" (UID: \"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.218711 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvlr9\" (UniqueName: \"kubernetes.io/projected/0969458b-57a6-4ec6-92b9-870ee1cfc6da-kube-api-access-lvlr9\") pod \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\" (UID: \"0969458b-57a6-4ec6-92b9-870ee1cfc6da\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.219453 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" (UID: "f7f3b5a2-58d0-45cc-9547-8b9eea1a9040"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.219549 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db4c80f2-bd41-4e94-b8e9-c8d6616357e6" (UID: "db4c80f2-bd41-4e94-b8e9-c8d6616357e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.219607 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0969458b-57a6-4ec6-92b9-870ee1cfc6da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0969458b-57a6-4ec6-92b9-870ee1cfc6da" (UID: "0969458b-57a6-4ec6-92b9-870ee1cfc6da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.224137 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-kube-api-access-95mdf" (OuterVolumeSpecName: "kube-api-access-95mdf") pod "db4c80f2-bd41-4e94-b8e9-c8d6616357e6" (UID: "db4c80f2-bd41-4e94-b8e9-c8d6616357e6"). InnerVolumeSpecName "kube-api-access-95mdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.224268 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0969458b-57a6-4ec6-92b9-870ee1cfc6da-kube-api-access-lvlr9" (OuterVolumeSpecName: "kube-api-access-lvlr9") pod "0969458b-57a6-4ec6-92b9-870ee1cfc6da" (UID: "0969458b-57a6-4ec6-92b9-870ee1cfc6da"). InnerVolumeSpecName "kube-api-access-lvlr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.224546 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-kube-api-access-dflbc" (OuterVolumeSpecName: "kube-api-access-dflbc") pod "f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" (UID: "f7f3b5a2-58d0-45cc-9547-8b9eea1a9040"). InnerVolumeSpecName "kube-api-access-dflbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.320525 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng4bf\" (UniqueName: \"kubernetes.io/projected/595fb7b3-d70c-4cf2-98c4-287fff3ec997-kube-api-access-ng4bf\") pod \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.320687 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6wzq\" (UniqueName: \"kubernetes.io/projected/c5b4081a-2347-45b3-b5db-07de3d5a90c2-kube-api-access-x6wzq\") pod \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.320755 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5b4081a-2347-45b3-b5db-07de3d5a90c2-operator-scripts\") pod \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\" (UID: \"c5b4081a-2347-45b3-b5db-07de3d5a90c2\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.320802 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/595fb7b3-d70c-4cf2-98c4-287fff3ec997-operator-scripts\") pod \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\" (UID: \"595fb7b3-d70c-4cf2-98c4-287fff3ec997\") " Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321240 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95mdf\" (UniqueName: \"kubernetes.io/projected/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-kube-api-access-95mdf\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321265 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0969458b-57a6-4ec6-92b9-870ee1cfc6da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321275 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db4c80f2-bd41-4e94-b8e9-c8d6616357e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321286 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321297 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvlr9\" (UniqueName: \"kubernetes.io/projected/0969458b-57a6-4ec6-92b9-870ee1cfc6da-kube-api-access-lvlr9\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321331 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dflbc\" (UniqueName: \"kubernetes.io/projected/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040-kube-api-access-dflbc\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.321950 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5b4081a-2347-45b3-b5db-07de3d5a90c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5b4081a-2347-45b3-b5db-07de3d5a90c2" (UID: "c5b4081a-2347-45b3-b5db-07de3d5a90c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.322002 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/595fb7b3-d70c-4cf2-98c4-287fff3ec997-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "595fb7b3-d70c-4cf2-98c4-287fff3ec997" (UID: "595fb7b3-d70c-4cf2-98c4-287fff3ec997"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.324373 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595fb7b3-d70c-4cf2-98c4-287fff3ec997-kube-api-access-ng4bf" (OuterVolumeSpecName: "kube-api-access-ng4bf") pod "595fb7b3-d70c-4cf2-98c4-287fff3ec997" (UID: "595fb7b3-d70c-4cf2-98c4-287fff3ec997"). InnerVolumeSpecName "kube-api-access-ng4bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.324943 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5b4081a-2347-45b3-b5db-07de3d5a90c2-kube-api-access-x6wzq" (OuterVolumeSpecName: "kube-api-access-x6wzq") pod "c5b4081a-2347-45b3-b5db-07de3d5a90c2" (UID: "c5b4081a-2347-45b3-b5db-07de3d5a90c2"). InnerVolumeSpecName "kube-api-access-x6wzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.423181 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6wzq\" (UniqueName: \"kubernetes.io/projected/c5b4081a-2347-45b3-b5db-07de3d5a90c2-kube-api-access-x6wzq\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.423226 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5b4081a-2347-45b3-b5db-07de3d5a90c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.423244 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/595fb7b3-d70c-4cf2-98c4-287fff3ec997-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.423261 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng4bf\" (UniqueName: \"kubernetes.io/projected/595fb7b3-d70c-4cf2-98c4-287fff3ec997-kube-api-access-ng4bf\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.843216 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-46e9-account-create-update-dbkf8" event={"ID":"db4c80f2-bd41-4e94-b8e9-c8d6616357e6","Type":"ContainerDied","Data":"039f0079010093bdcb3ab0b055ef144ad6edc7c23e1f94e75dfb63cf0c75f46f"} Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.843254 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="039f0079010093bdcb3ab0b055ef144ad6edc7c23e1f94e75dfb63cf0c75f46f" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.843268 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-46e9-account-create-update-dbkf8" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.844722 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b3a9-account-create-update-jg27q" event={"ID":"595fb7b3-d70c-4cf2-98c4-287fff3ec997","Type":"ContainerDied","Data":"7bf7ac41a8d5ba915f76e2851b76d8525cddab58be9db462638f15e16117c5f8"} Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.844746 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bf7ac41a8d5ba915f76e2851b76d8525cddab58be9db462638f15e16117c5f8" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.844792 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b3a9-account-create-update-jg27q" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.858195 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dk9q8" event={"ID":"9f9bcb09-6b18-4357-b474-0db2b33e1641","Type":"ContainerStarted","Data":"c100dfeef4fdb34a70898323eb438a18537e6364c365397f0168d3e0197551f1"} Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.861406 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ctlv8" event={"ID":"f7f3b5a2-58d0-45cc-9547-8b9eea1a9040","Type":"ContainerDied","Data":"e969d76211bbcf8a1bb287f5680b90f666fe4d1fec14238eac2f0bffb77d9a83"} Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.861444 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e969d76211bbcf8a1bb287f5680b90f666fe4d1fec14238eac2f0bffb77d9a83" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.861482 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ctlv8" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.870466 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e045-account-create-update-c77tp" event={"ID":"c5b4081a-2347-45b3-b5db-07de3d5a90c2","Type":"ContainerDied","Data":"0aad27b66df05b49626774a793934d00186dc98c5c7f10494eaca54ca206e645"} Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.870535 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0aad27b66df05b49626774a793934d00186dc98c5c7f10494eaca54ca206e645" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.870511 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e045-account-create-update-c77tp" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.872352 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qbg4x" event={"ID":"0969458b-57a6-4ec6-92b9-870ee1cfc6da","Type":"ContainerDied","Data":"c9dd3f4fce58c702db25407343befe5f0385c5ddad4499076a3d1b6b16807812"} Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.872392 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9dd3f4fce58c702db25407343befe5f0385c5ddad4499076a3d1b6b16807812" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.872449 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qbg4x" Dec 02 10:38:10 crc kubenswrapper[4679]: I1202 10:38:10.906836 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-dk9q8" podStartSLOduration=9.675393477 podStartE2EDuration="14.906813803s" podCreationTimestamp="2025-12-02 10:37:56 +0000 UTC" firstStartedPulling="2025-12-02 10:38:04.715069816 +0000 UTC m=+1138.045208676" lastFinishedPulling="2025-12-02 10:38:09.946490142 +0000 UTC m=+1143.276629002" observedRunningTime="2025-12-02 10:38:10.899366869 +0000 UTC m=+1144.229505809" watchObservedRunningTime="2025-12-02 10:38:10.906813803 +0000 UTC m=+1144.236952673" Dec 02 10:38:13 crc kubenswrapper[4679]: I1202 10:38:13.899794 4679 generic.go:334] "Generic (PLEG): container finished" podID="9f9bcb09-6b18-4357-b474-0db2b33e1641" containerID="c100dfeef4fdb34a70898323eb438a18537e6364c365397f0168d3e0197551f1" exitCode=0 Dec 02 10:38:13 crc kubenswrapper[4679]: I1202 10:38:13.899889 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dk9q8" event={"ID":"9f9bcb09-6b18-4357-b474-0db2b33e1641","Type":"ContainerDied","Data":"c100dfeef4fdb34a70898323eb438a18537e6364c365397f0168d3e0197551f1"} Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.243924 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.362616 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.406204 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-config-data\") pod \"9f9bcb09-6b18-4357-b474-0db2b33e1641\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.406277 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krtgm\" (UniqueName: \"kubernetes.io/projected/9f9bcb09-6b18-4357-b474-0db2b33e1641-kube-api-access-krtgm\") pod \"9f9bcb09-6b18-4357-b474-0db2b33e1641\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.406348 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-combined-ca-bundle\") pod \"9f9bcb09-6b18-4357-b474-0db2b33e1641\" (UID: \"9f9bcb09-6b18-4357-b474-0db2b33e1641\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.412337 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmlpw"] Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.412832 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qmlpw" podUID="08d8ff70-7182-405f-b253-96f212f13410" containerName="dnsmasq-dns" containerID="cri-o://563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b" gracePeriod=10 Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.417571 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f9bcb09-6b18-4357-b474-0db2b33e1641-kube-api-access-krtgm" (OuterVolumeSpecName: "kube-api-access-krtgm") pod "9f9bcb09-6b18-4357-b474-0db2b33e1641" (UID: "9f9bcb09-6b18-4357-b474-0db2b33e1641"). InnerVolumeSpecName "kube-api-access-krtgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.465534 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f9bcb09-6b18-4357-b474-0db2b33e1641" (UID: "9f9bcb09-6b18-4357-b474-0db2b33e1641"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.504523 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-config-data" (OuterVolumeSpecName: "config-data") pod "9f9bcb09-6b18-4357-b474-0db2b33e1641" (UID: "9f9bcb09-6b18-4357-b474-0db2b33e1641"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.509038 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.509086 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krtgm\" (UniqueName: \"kubernetes.io/projected/9f9bcb09-6b18-4357-b474-0db2b33e1641-kube-api-access-krtgm\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.509101 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f9bcb09-6b18-4357-b474-0db2b33e1641-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.776617 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.914909 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-nb\") pod \"08d8ff70-7182-405f-b253-96f212f13410\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.915062 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-dns-svc\") pod \"08d8ff70-7182-405f-b253-96f212f13410\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.915283 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6dm2\" (UniqueName: \"kubernetes.io/projected/08d8ff70-7182-405f-b253-96f212f13410-kube-api-access-c6dm2\") pod \"08d8ff70-7182-405f-b253-96f212f13410\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.915437 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-sb\") pod \"08d8ff70-7182-405f-b253-96f212f13410\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.915538 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-config\") pod \"08d8ff70-7182-405f-b253-96f212f13410\" (UID: \"08d8ff70-7182-405f-b253-96f212f13410\") " Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.919788 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d8ff70-7182-405f-b253-96f212f13410-kube-api-access-c6dm2" (OuterVolumeSpecName: "kube-api-access-c6dm2") pod "08d8ff70-7182-405f-b253-96f212f13410" (UID: "08d8ff70-7182-405f-b253-96f212f13410"). InnerVolumeSpecName "kube-api-access-c6dm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.939121 4679 generic.go:334] "Generic (PLEG): container finished" podID="08d8ff70-7182-405f-b253-96f212f13410" containerID="563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b" exitCode=0 Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.939534 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmlpw" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.939449 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmlpw" event={"ID":"08d8ff70-7182-405f-b253-96f212f13410","Type":"ContainerDied","Data":"563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b"} Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.939746 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmlpw" event={"ID":"08d8ff70-7182-405f-b253-96f212f13410","Type":"ContainerDied","Data":"f76cec11eed3aac7e769da7a5e6b1cb25b72e9a53fa477d49ca227359625f32a"} Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.939791 4679 scope.go:117] "RemoveContainer" containerID="563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.942911 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dk9q8" event={"ID":"9f9bcb09-6b18-4357-b474-0db2b33e1641","Type":"ContainerDied","Data":"3b406f25b4ee23e589aab395dfc2adf230fd7dbde0bf8d20cf5e35b360d6ff7d"} Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.942957 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b406f25b4ee23e589aab395dfc2adf230fd7dbde0bf8d20cf5e35b360d6ff7d" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.943019 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dk9q8" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.966611 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-config" (OuterVolumeSpecName: "config") pod "08d8ff70-7182-405f-b253-96f212f13410" (UID: "08d8ff70-7182-405f-b253-96f212f13410"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.970324 4679 scope.go:117] "RemoveContainer" containerID="3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.976872 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "08d8ff70-7182-405f-b253-96f212f13410" (UID: "08d8ff70-7182-405f-b253-96f212f13410"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.984844 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08d8ff70-7182-405f-b253-96f212f13410" (UID: "08d8ff70-7182-405f-b253-96f212f13410"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.993919 4679 scope.go:117] "RemoveContainer" containerID="563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.994132 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08d8ff70-7182-405f-b253-96f212f13410" (UID: "08d8ff70-7182-405f-b253-96f212f13410"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:15 crc kubenswrapper[4679]: E1202 10:38:15.998970 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b\": container with ID starting with 563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b not found: ID does not exist" containerID="563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.999048 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b"} err="failed to get container status \"563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b\": rpc error: code = NotFound desc = could not find container \"563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b\": container with ID starting with 563b202ad5a3303d18651a3c0e699d884cd5dee5a210ef51c953ec3b26da936b not found: ID does not exist" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.999100 4679 scope.go:117] "RemoveContainer" containerID="3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747" Dec 02 10:38:15 crc kubenswrapper[4679]: E1202 10:38:15.999566 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747\": container with ID starting with 3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747 not found: ID does not exist" containerID="3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747" Dec 02 10:38:15 crc kubenswrapper[4679]: I1202 10:38:15.999611 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747"} err="failed to get container status \"3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747\": rpc error: code = NotFound desc = could not find container \"3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747\": container with ID starting with 3d39cbc28d950b8cdd6c7fe75f0e33df2d237976bd2565a63de721e6dbe1c747 not found: ID does not exist" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.017664 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6dm2\" (UniqueName: \"kubernetes.io/projected/08d8ff70-7182-405f-b253-96f212f13410-kube-api-access-c6dm2\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.017740 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.017757 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.017769 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.017779 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08d8ff70-7182-405f-b253-96f212f13410-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168267 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-9vhtf"] Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168648 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595fb7b3-d70c-4cf2-98c4-287fff3ec997" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168665 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="595fb7b3-d70c-4cf2-98c4-287fff3ec997" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168685 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d8ff70-7182-405f-b253-96f212f13410" containerName="init" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168691 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d8ff70-7182-405f-b253-96f212f13410" containerName="init" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168700 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f9bcb09-6b18-4357-b474-0db2b33e1641" containerName="keystone-db-sync" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168705 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f9bcb09-6b18-4357-b474-0db2b33e1641" containerName="keystone-db-sync" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168713 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0969458b-57a6-4ec6-92b9-870ee1cfc6da" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168720 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0969458b-57a6-4ec6-92b9-870ee1cfc6da" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168730 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168737 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168748 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4c80f2-bd41-4e94-b8e9-c8d6616357e6" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168753 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4c80f2-bd41-4e94-b8e9-c8d6616357e6" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168769 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b4081a-2347-45b3-b5db-07de3d5a90c2" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168775 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b4081a-2347-45b3-b5db-07de3d5a90c2" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168785 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d8ff70-7182-405f-b253-96f212f13410" containerName="dnsmasq-dns" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168791 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d8ff70-7182-405f-b253-96f212f13410" containerName="dnsmasq-dns" Dec 02 10:38:16 crc kubenswrapper[4679]: E1202 10:38:16.168800 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d57d916-49db-43e8-b1cd-7bd5406d4812" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168805 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d57d916-49db-43e8-b1cd-7bd5406d4812" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168942 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0969458b-57a6-4ec6-92b9-870ee1cfc6da" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168955 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d8ff70-7182-405f-b253-96f212f13410" containerName="dnsmasq-dns" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168967 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="595fb7b3-d70c-4cf2-98c4-287fff3ec997" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168975 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4c80f2-bd41-4e94-b8e9-c8d6616357e6" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168984 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d57d916-49db-43e8-b1cd-7bd5406d4812" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.168996 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" containerName="mariadb-database-create" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.169001 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5b4081a-2347-45b3-b5db-07de3d5a90c2" containerName="mariadb-account-create-update" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.169016 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f9bcb09-6b18-4357-b474-0db2b33e1641" containerName="keystone-db-sync" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.169843 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.187731 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-9vhtf"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.222026 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rf2ql"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.223018 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.225832 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5x6pg" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.226114 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.226289 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.226457 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.239577 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.259418 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rf2ql"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.305292 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmlpw"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323729 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-config\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323802 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-combined-ca-bundle\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323834 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323856 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-credential-keys\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323908 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-fernet-keys\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323931 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-224p2\" (UniqueName: \"kubernetes.io/projected/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-kube-api-access-224p2\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323959 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323981 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzrwk\" (UniqueName: \"kubernetes.io/projected/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-kube-api-access-fzrwk\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.323925 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmlpw"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.324069 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.324147 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-config-data\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.324193 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-scripts\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.324299 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-svc\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.396502 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-w58bz"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.398014 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.401676 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.402015 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9bb5d" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.403940 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9b9b55977-kp8vd"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.405426 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.405988 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425421 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-config\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425505 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-combined-ca-bundle\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425545 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425565 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-credential-keys\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425602 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-fernet-keys\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425619 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-224p2\" (UniqueName: \"kubernetes.io/projected/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-kube-api-access-224p2\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425638 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425661 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzrwk\" (UniqueName: \"kubernetes.io/projected/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-kube-api-access-fzrwk\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425684 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425708 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-config-data\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425729 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-scripts\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.425777 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-svc\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.426702 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-svc\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.427053 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-config\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.427332 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.428289 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.432766 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.433636 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.433842 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.434389 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-combined-ca-bundle\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.436967 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.458053 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9b9b55977-kp8vd"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.462896 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-w58bz"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.471669 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-pzdt5" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.480349 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-bqb7p"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.500665 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-credential-keys\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.500784 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-scripts\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.501123 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-fernet-keys\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.501421 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzrwk\" (UniqueName: \"kubernetes.io/projected/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-kube-api-access-fzrwk\") pod \"dnsmasq-dns-55fff446b9-9vhtf\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.504852 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.513438 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-config-data\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.514180 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pt7cd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.514886 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.525034 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-224p2\" (UniqueName: \"kubernetes.io/projected/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-kube-api-access-224p2\") pod \"keystone-bootstrap-rf2ql\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528506 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-combined-ca-bundle\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528553 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-scripts\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528578 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-config\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528614 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzq2t\" (UniqueName: \"kubernetes.io/projected/b3e30d8f-75cb-474d-85b0-6656b987dae8-kube-api-access-rzq2t\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528639 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3e30d8f-75cb-474d-85b0-6656b987dae8-horizon-secret-key\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528664 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-config-data\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.528935 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdwxz\" (UniqueName: \"kubernetes.io/projected/e1284622-943e-4f14-b59c-79eb084e8ea5-kube-api-access-pdwxz\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.529148 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3e30d8f-75cb-474d-85b0-6656b987dae8-logs\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.537527 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.549934 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bqb7p"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.564290 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.564593 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.571347 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.576147 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.616567 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.618648 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.618849 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632495 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdwxz\" (UniqueName: \"kubernetes.io/projected/e1284622-943e-4f14-b59c-79eb084e8ea5-kube-api-access-pdwxz\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632558 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-combined-ca-bundle\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632601 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-config-data\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632624 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3e30d8f-75cb-474d-85b0-6656b987dae8-logs\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632648 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-scripts\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632663 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-combined-ca-bundle\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632678 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/048db06f-54db-4862-b8c8-b7e9cbb55095-etc-machine-id\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632695 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-scripts\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632718 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-config\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632747 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzq2t\" (UniqueName: \"kubernetes.io/projected/b3e30d8f-75cb-474d-85b0-6656b987dae8-kube-api-access-rzq2t\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632765 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3e30d8f-75cb-474d-85b0-6656b987dae8-horizon-secret-key\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632780 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-db-sync-config-data\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632802 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-config-data\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.632818 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ptzq\" (UniqueName: \"kubernetes.io/projected/048db06f-54db-4862-b8c8-b7e9cbb55095-kube-api-access-8ptzq\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.633918 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-scripts\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.634173 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3e30d8f-75cb-474d-85b0-6656b987dae8-logs\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.637892 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-9kjcb"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.638824 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-config-data\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.639449 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.639824 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-config\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.644735 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-combined-ca-bundle\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.672059 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.674665 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.674896 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-jqtrr" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.676753 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3e30d8f-75cb-474d-85b0-6656b987dae8-horizon-secret-key\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.684000 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdwxz\" (UniqueName: \"kubernetes.io/projected/e1284622-943e-4f14-b59c-79eb084e8ea5-kube-api-access-pdwxz\") pod \"neutron-db-sync-w58bz\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.691077 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzq2t\" (UniqueName: \"kubernetes.io/projected/b3e30d8f-75cb-474d-85b0-6656b987dae8-kube-api-access-rzq2t\") pod \"horizon-9b9b55977-kp8vd\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.718147 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-9kjcb"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.726662 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.735138 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/055921c7-b108-40a9-a4da-2947ca27c2d8-logs\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.735336 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnxjd\" (UniqueName: \"kubernetes.io/projected/055921c7-b108-40a9-a4da-2947ca27c2d8-kube-api-access-qnxjd\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.735478 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-config-data\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.735698 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-config-data\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.735867 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-scripts\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.735989 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-combined-ca-bundle\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.736370 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.736535 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-config-data\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.736675 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-log-httpd\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.736802 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-scripts\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.743399 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-combined-ca-bundle\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.744749 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-scripts\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.744854 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/048db06f-54db-4862-b8c8-b7e9cbb55095-etc-machine-id\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.744966 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-config-data\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745123 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-combined-ca-bundle\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745209 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745285 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-run-httpd\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745441 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-db-sync-config-data\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745595 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb9w6\" (UniqueName: \"kubernetes.io/projected/470284bf-e122-400f-9722-f8d4f23a75cf-kube-api-access-hb9w6\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745794 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ptzq\" (UniqueName: \"kubernetes.io/projected/048db06f-54db-4862-b8c8-b7e9cbb55095-kube-api-access-8ptzq\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.745324 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/048db06f-54db-4862-b8c8-b7e9cbb55095-etc-machine-id\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.747604 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-9vhtf"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.748025 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-scripts\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.748850 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-db-sync-config-data\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.800912 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jtqff"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.802504 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.815580 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ptzq\" (UniqueName: \"kubernetes.io/projected/048db06f-54db-4862-b8c8-b7e9cbb55095-kube-api-access-8ptzq\") pod \"cinder-db-sync-bqb7p\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847016 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-log-httpd\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847057 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-scripts\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847105 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-combined-ca-bundle\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847122 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847136 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-run-httpd\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847162 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb9w6\" (UniqueName: \"kubernetes.io/projected/470284bf-e122-400f-9722-f8d4f23a75cf-kube-api-access-hb9w6\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847193 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/055921c7-b108-40a9-a4da-2947ca27c2d8-logs\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847206 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnxjd\" (UniqueName: \"kubernetes.io/projected/055921c7-b108-40a9-a4da-2947ca27c2d8-kube-api-access-qnxjd\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847237 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-config-data\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847258 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-config-data\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847282 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-scripts\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.847353 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.848375 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-log-httpd\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.854812 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/055921c7-b108-40a9-a4da-2947ca27c2d8-logs\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.855177 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-run-httpd\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.866294 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-config-data\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.869811 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jtqff"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.881493 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-combined-ca-bundle\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.887383 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5447f744df-7bx4r"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.889147 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.890044 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-config-data\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.890988 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-scripts\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.897008 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-scripts\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.897908 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.898372 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.924782 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.929042 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnxjd\" (UniqueName: \"kubernetes.io/projected/055921c7-b108-40a9-a4da-2947ca27c2d8-kube-api-access-qnxjd\") pod \"placement-db-sync-9kjcb\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.929083 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb9w6\" (UniqueName: \"kubernetes.io/projected/470284bf-e122-400f-9722-f8d4f23a75cf-kube-api-access-hb9w6\") pod \"ceilometer-0\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " pod="openstack/ceilometer-0" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.953371 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-config\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.953430 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.953538 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.953584 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzxwl\" (UniqueName: \"kubernetes.io/projected/78457c3d-d7fe-413c-aca4-63a57deec717-kube-api-access-lzxwl\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.953660 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.953738 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.959169 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d8ff70-7182-405f-b253-96f212f13410" path="/var/lib/kubelet/pods/08d8ff70-7182-405f-b253-96f212f13410/volumes" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.970757 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5447f744df-7bx4r"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.977788 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-86lxg"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.978768 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.983021 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.983223 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b8zhd" Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.994271 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-86lxg"] Dec 02 10:38:16 crc kubenswrapper[4679]: I1202 10:38:16.994740 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.034979 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.048132 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9kjcb" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.068338 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzxwl\" (UniqueName: \"kubernetes.io/projected/78457c3d-d7fe-413c-aca4-63a57deec717-kube-api-access-lzxwl\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.068451 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-logs\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.068537 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pptfr\" (UniqueName: \"kubernetes.io/projected/f071f49e-c3e2-4919-a930-dd5fd654cf55-kube-api-access-pptfr\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069145 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-db-sync-config-data\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069179 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069209 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-horizon-secret-key\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069299 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-scripts\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069415 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-config-data\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069450 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069481 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2c95\" (UniqueName: \"kubernetes.io/projected/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-kube-api-access-c2c95\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069509 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-combined-ca-bundle\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069566 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-config\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069604 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.069736 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.071722 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-config\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.073087 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.073435 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.074530 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.074687 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.098078 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzxwl\" (UniqueName: \"kubernetes.io/projected/78457c3d-d7fe-413c-aca4-63a57deec717-kube-api-access-lzxwl\") pod \"dnsmasq-dns-76fcf4b695-jtqff\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.170807 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-config-data\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.170871 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2c95\" (UniqueName: \"kubernetes.io/projected/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-kube-api-access-c2c95\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.170894 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-combined-ca-bundle\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.170988 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-logs\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.171018 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pptfr\" (UniqueName: \"kubernetes.io/projected/f071f49e-c3e2-4919-a930-dd5fd654cf55-kube-api-access-pptfr\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.171043 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-db-sync-config-data\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.171063 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-horizon-secret-key\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.171106 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-scripts\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.172035 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-logs\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.172278 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-scripts\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.173037 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-config-data\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.177246 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-horizon-secret-key\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.182845 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-combined-ca-bundle\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.183238 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-db-sync-config-data\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.193206 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2c95\" (UniqueName: \"kubernetes.io/projected/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-kube-api-access-c2c95\") pod \"horizon-5447f744df-7bx4r\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.196790 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pptfr\" (UniqueName: \"kubernetes.io/projected/f071f49e-c3e2-4919-a930-dd5fd654cf55-kube-api-access-pptfr\") pod \"barbican-db-sync-86lxg\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.289281 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.322961 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.338218 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-86lxg" Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.355937 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rf2ql"] Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.563445 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-w58bz"] Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.606226 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9b9b55977-kp8vd"] Dec 02 10:38:17 crc kubenswrapper[4679]: W1202 10:38:17.618022 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3e30d8f_75cb_474d_85b0_6656b987dae8.slice/crio-bd11339b1cf56e2b1b79430cd0e9f966b664c361ab70a9a964cc0358a825beb3 WatchSource:0}: Error finding container bd11339b1cf56e2b1b79430cd0e9f966b664c361ab70a9a964cc0358a825beb3: Status 404 returned error can't find the container with id bd11339b1cf56e2b1b79430cd0e9f966b664c361ab70a9a964cc0358a825beb3 Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.624669 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-9vhtf"] Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.766388 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:38:17 crc kubenswrapper[4679]: W1202 10:38:17.780833 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod048db06f_54db_4862_b8c8_b7e9cbb55095.slice/crio-85fda557c9826c0ec018e32ca57299f707fbc6fa95cac2e07570b25f88150d2c WatchSource:0}: Error finding container 85fda557c9826c0ec018e32ca57299f707fbc6fa95cac2e07570b25f88150d2c: Status 404 returned error can't find the container with id 85fda557c9826c0ec018e32ca57299f707fbc6fa95cac2e07570b25f88150d2c Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.785378 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-bqb7p"] Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.981404 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-9kjcb"] Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.993647 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" event={"ID":"1438472e-8d3d-4a6a-a3fd-7b02bc501e06","Type":"ContainerStarted","Data":"7163a599bfa0f4b7691e72eaeb75542afaa305a0787e6f4848630dc62d725638"} Dec 02 10:38:17 crc kubenswrapper[4679]: I1202 10:38:17.995402 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-w58bz" event={"ID":"e1284622-943e-4f14-b59c-79eb084e8ea5","Type":"ContainerStarted","Data":"1b1111d9d3d4e2aefa4c2a8224e76665ca258e1f9aedd5e79be07583962674ec"} Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.005960 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"470284bf-e122-400f-9722-f8d4f23a75cf","Type":"ContainerStarted","Data":"a1101a1e7d2098618b2e9c56637858ec4b1233f58163026c7b714b4f4c4577bf"} Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.008244 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9kjcb" event={"ID":"055921c7-b108-40a9-a4da-2947ca27c2d8","Type":"ContainerStarted","Data":"c0305d18386fa0d9f6702fbe82910fca2bec0b322a2a701a05648bf27a72a4a2"} Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.010849 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rf2ql" event={"ID":"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3","Type":"ContainerStarted","Data":"8e12ea3a1bb7512e0d4f1e073ffc58e884d2c83a5e0978d3630bb44d1107dec3"} Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.012236 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqb7p" event={"ID":"048db06f-54db-4862-b8c8-b7e9cbb55095","Type":"ContainerStarted","Data":"85fda557c9826c0ec018e32ca57299f707fbc6fa95cac2e07570b25f88150d2c"} Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.013886 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9b9b55977-kp8vd" event={"ID":"b3e30d8f-75cb-474d-85b0-6656b987dae8","Type":"ContainerStarted","Data":"bd11339b1cf56e2b1b79430cd0e9f966b664c361ab70a9a964cc0358a825beb3"} Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.044763 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rf2ql" podStartSLOduration=2.044737545 podStartE2EDuration="2.044737545s" podCreationTimestamp="2025-12-02 10:38:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:18.036132127 +0000 UTC m=+1151.366271017" watchObservedRunningTime="2025-12-02 10:38:18.044737545 +0000 UTC m=+1151.374876405" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.076279 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jtqff"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.102256 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5447f744df-7bx4r"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.110001 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-86lxg"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.313760 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5447f744df-7bx4r"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.328686 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.361369 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-794d9b8d7f-pr9jf"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.364226 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.397695 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-794d9b8d7f-pr9jf"] Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.406546 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a750d67-054d-45d8-b1bb-30d89a141d72-logs\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.406609 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-scripts\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.406667 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a750d67-054d-45d8-b1bb-30d89a141d72-horizon-secret-key\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.406734 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c29f7\" (UniqueName: \"kubernetes.io/projected/6a750d67-054d-45d8-b1bb-30d89a141d72-kube-api-access-c29f7\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.406772 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-config-data\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.508291 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c29f7\" (UniqueName: \"kubernetes.io/projected/6a750d67-054d-45d8-b1bb-30d89a141d72-kube-api-access-c29f7\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.508940 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-config-data\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.509042 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a750d67-054d-45d8-b1bb-30d89a141d72-logs\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.509086 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-scripts\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.509143 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a750d67-054d-45d8-b1bb-30d89a141d72-horizon-secret-key\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.515655 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-config-data\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.518699 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a750d67-054d-45d8-b1bb-30d89a141d72-logs\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.518907 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a750d67-054d-45d8-b1bb-30d89a141d72-horizon-secret-key\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.520032 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-scripts\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.533099 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c29f7\" (UniqueName: \"kubernetes.io/projected/6a750d67-054d-45d8-b1bb-30d89a141d72-kube-api-access-c29f7\") pod \"horizon-794d9b8d7f-pr9jf\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:18 crc kubenswrapper[4679]: I1202 10:38:18.712231 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.031566 4679 generic.go:334] "Generic (PLEG): container finished" podID="1438472e-8d3d-4a6a-a3fd-7b02bc501e06" containerID="6c60801c062219c046b4eddcbf3c61beb5316ae88b800a4b1c34c8f4b239c391" exitCode=0 Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.031682 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" event={"ID":"1438472e-8d3d-4a6a-a3fd-7b02bc501e06","Type":"ContainerDied","Data":"6c60801c062219c046b4eddcbf3c61beb5316ae88b800a4b1c34c8f4b239c391"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.035348 4679 generic.go:334] "Generic (PLEG): container finished" podID="78457c3d-d7fe-413c-aca4-63a57deec717" containerID="3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780" exitCode=0 Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.035417 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" event={"ID":"78457c3d-d7fe-413c-aca4-63a57deec717","Type":"ContainerDied","Data":"3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.035444 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" event={"ID":"78457c3d-d7fe-413c-aca4-63a57deec717","Type":"ContainerStarted","Data":"27d6a3d032f67c7079cecbd807805066d8af5043e67eea980826575cb7272ce2"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.065140 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-w58bz" event={"ID":"e1284622-943e-4f14-b59c-79eb084e8ea5","Type":"ContainerStarted","Data":"50450f1a8db17111a960666ee3f2b7e9231d0580c967b8b19c101e1179dc09a0"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.079132 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rf2ql" event={"ID":"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3","Type":"ContainerStarted","Data":"81917a8d774a490d5fa6c3c95d1d994ec418a92049e6aeaafe5bf801b6794d0f"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.082922 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-86lxg" event={"ID":"f071f49e-c3e2-4919-a930-dd5fd654cf55","Type":"ContainerStarted","Data":"602b714ce0dc4f36eaf21508a9a0d1b10fc2be668769576e7d64ab13a3b1d1d8"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.092230 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5447f744df-7bx4r" event={"ID":"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6","Type":"ContainerStarted","Data":"cac18f4e70c3b8e8c2405e713d3809adea65a56e9065192c35edbeb7cd1f1c43"} Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.132392 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-w58bz" podStartSLOduration=3.13236894 podStartE2EDuration="3.13236894s" podCreationTimestamp="2025-12-02 10:38:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:19.101656146 +0000 UTC m=+1152.431795006" watchObservedRunningTime="2025-12-02 10:38:19.13236894 +0000 UTC m=+1152.462507800" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.177881 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-794d9b8d7f-pr9jf"] Dec 02 10:38:19 crc kubenswrapper[4679]: W1202 10:38:19.206584 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a750d67_054d_45d8_b1bb_30d89a141d72.slice/crio-0521018c5eda668b1efddc8e3b13052a50d23d58e3e3de60b73e6962d063c77e WatchSource:0}: Error finding container 0521018c5eda668b1efddc8e3b13052a50d23d58e3e3de60b73e6962d063c77e: Status 404 returned error can't find the container with id 0521018c5eda668b1efddc8e3b13052a50d23d58e3e3de60b73e6962d063c77e Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.527794 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.631485 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-svc\") pod \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.631540 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-config\") pod \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.631595 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzrwk\" (UniqueName: \"kubernetes.io/projected/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-kube-api-access-fzrwk\") pod \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.631716 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-nb\") pod \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.631790 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-sb\") pod \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.631847 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-swift-storage-0\") pod \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\" (UID: \"1438472e-8d3d-4a6a-a3fd-7b02bc501e06\") " Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.657121 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-kube-api-access-fzrwk" (OuterVolumeSpecName: "kube-api-access-fzrwk") pod "1438472e-8d3d-4a6a-a3fd-7b02bc501e06" (UID: "1438472e-8d3d-4a6a-a3fd-7b02bc501e06"). InnerVolumeSpecName "kube-api-access-fzrwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.663724 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1438472e-8d3d-4a6a-a3fd-7b02bc501e06" (UID: "1438472e-8d3d-4a6a-a3fd-7b02bc501e06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.671178 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-config" (OuterVolumeSpecName: "config") pod "1438472e-8d3d-4a6a-a3fd-7b02bc501e06" (UID: "1438472e-8d3d-4a6a-a3fd-7b02bc501e06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.679683 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1438472e-8d3d-4a6a-a3fd-7b02bc501e06" (UID: "1438472e-8d3d-4a6a-a3fd-7b02bc501e06"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.695165 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1438472e-8d3d-4a6a-a3fd-7b02bc501e06" (UID: "1438472e-8d3d-4a6a-a3fd-7b02bc501e06"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.695786 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1438472e-8d3d-4a6a-a3fd-7b02bc501e06" (UID: "1438472e-8d3d-4a6a-a3fd-7b02bc501e06"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.733677 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzrwk\" (UniqueName: \"kubernetes.io/projected/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-kube-api-access-fzrwk\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.733721 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.733730 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.733740 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.733751 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:19 crc kubenswrapper[4679]: I1202 10:38:19.733760 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438472e-8d3d-4a6a-a3fd-7b02bc501e06-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.121974 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" event={"ID":"1438472e-8d3d-4a6a-a3fd-7b02bc501e06","Type":"ContainerDied","Data":"7163a599bfa0f4b7691e72eaeb75542afaa305a0787e6f4848630dc62d725638"} Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.122062 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-9vhtf" Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.122300 4679 scope.go:117] "RemoveContainer" containerID="6c60801c062219c046b4eddcbf3c61beb5316ae88b800a4b1c34c8f4b239c391" Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.141516 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" event={"ID":"78457c3d-d7fe-413c-aca4-63a57deec717","Type":"ContainerStarted","Data":"5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9"} Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.141957 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.145864 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-794d9b8d7f-pr9jf" event={"ID":"6a750d67-054d-45d8-b1bb-30d89a141d72","Type":"ContainerStarted","Data":"0521018c5eda668b1efddc8e3b13052a50d23d58e3e3de60b73e6962d063c77e"} Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.196157 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-9vhtf"] Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.202973 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-9vhtf"] Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.209796 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" podStartSLOduration=4.209776381 podStartE2EDuration="4.209776381s" podCreationTimestamp="2025-12-02 10:38:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:20.194607614 +0000 UTC m=+1153.524746494" watchObservedRunningTime="2025-12-02 10:38:20.209776381 +0000 UTC m=+1153.539915241" Dec 02 10:38:20 crc kubenswrapper[4679]: I1202 10:38:20.920664 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1438472e-8d3d-4a6a-a3fd-7b02bc501e06" path="/var/lib/kubelet/pods/1438472e-8d3d-4a6a-a3fd-7b02bc501e06/volumes" Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.887938 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9b9b55977-kp8vd"] Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.941542 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6bd7d57448-84dtd"] Dec 02 10:38:24 crc kubenswrapper[4679]: E1202 10:38:24.941905 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1438472e-8d3d-4a6a-a3fd-7b02bc501e06" containerName="init" Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.941919 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1438472e-8d3d-4a6a-a3fd-7b02bc501e06" containerName="init" Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.942145 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="1438472e-8d3d-4a6a-a3fd-7b02bc501e06" containerName="init" Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.943205 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.946807 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 02 10:38:24 crc kubenswrapper[4679]: I1202 10:38:24.962674 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bd7d57448-84dtd"] Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.018173 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-794d9b8d7f-pr9jf"] Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.055645 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-combined-ca-bundle\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.055816 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-scripts\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.055882 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-config-data\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.056086 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-secret-key\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.056258 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-tls-certs\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.056384 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded5eaca-d45b-46bf-a770-50d909f0003e-logs\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.056632 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn297\" (UniqueName: \"kubernetes.io/projected/ded5eaca-d45b-46bf-a770-50d909f0003e-kube-api-access-bn297\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.069958 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6fd99d7d5b-sb96h"] Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.072863 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.078726 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fd99d7d5b-sb96h"] Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.182113 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn297\" (UniqueName: \"kubernetes.io/projected/ded5eaca-d45b-46bf-a770-50d909f0003e-kube-api-access-bn297\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.182459 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-combined-ca-bundle\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.182574 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4dd62da9-d481-4fc3-bbe6-108a54d98154-config-data\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.182675 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-combined-ca-bundle\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.182790 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-horizon-secret-key\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.182899 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-scripts\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.183022 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-config-data\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.183195 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-secret-key\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.184527 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-horizon-tls-certs\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.184696 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-tls-certs\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.184817 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded5eaca-d45b-46bf-a770-50d909f0003e-logs\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.184922 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chkvw\" (UniqueName: \"kubernetes.io/projected/4dd62da9-d481-4fc3-bbe6-108a54d98154-kube-api-access-chkvw\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.185035 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dd62da9-d481-4fc3-bbe6-108a54d98154-logs\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.185137 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4dd62da9-d481-4fc3-bbe6-108a54d98154-scripts\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.183701 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-scripts\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.184454 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-config-data\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.199496 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded5eaca-d45b-46bf-a770-50d909f0003e-logs\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.202623 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-secret-key\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.206043 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-tls-certs\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.216980 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-combined-ca-bundle\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.219086 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn297\" (UniqueName: \"kubernetes.io/projected/ded5eaca-d45b-46bf-a770-50d909f0003e-kube-api-access-bn297\") pod \"horizon-6bd7d57448-84dtd\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.237533 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zs6tz" event={"ID":"8d59898e-672f-4699-bf31-a964cd84913b","Type":"ContainerStarted","Data":"16c7e627e2b769fec6fded2a0edc40f16b7e1e272141289d48ed3bc7171d69c4"} Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.282622 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-zs6tz" podStartSLOduration=2.777599812 podStartE2EDuration="41.282605962s" podCreationTimestamp="2025-12-02 10:37:44 +0000 UTC" firstStartedPulling="2025-12-02 10:37:45.696579567 +0000 UTC m=+1119.026718427" lastFinishedPulling="2025-12-02 10:38:24.201585717 +0000 UTC m=+1157.531724577" observedRunningTime="2025-12-02 10:38:25.276778345 +0000 UTC m=+1158.606917205" watchObservedRunningTime="2025-12-02 10:38:25.282605962 +0000 UTC m=+1158.612744823" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286419 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-combined-ca-bundle\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286462 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4dd62da9-d481-4fc3-bbe6-108a54d98154-config-data\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286489 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-horizon-secret-key\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286577 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-horizon-tls-certs\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286636 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chkvw\" (UniqueName: \"kubernetes.io/projected/4dd62da9-d481-4fc3-bbe6-108a54d98154-kube-api-access-chkvw\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286670 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dd62da9-d481-4fc3-bbe6-108a54d98154-logs\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.286698 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4dd62da9-d481-4fc3-bbe6-108a54d98154-scripts\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.287502 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4dd62da9-d481-4fc3-bbe6-108a54d98154-scripts\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.287726 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4dd62da9-d481-4fc3-bbe6-108a54d98154-logs\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.288685 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4dd62da9-d481-4fc3-bbe6-108a54d98154-config-data\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.290346 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-horizon-secret-key\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.290913 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-combined-ca-bundle\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.291675 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4dd62da9-d481-4fc3-bbe6-108a54d98154-horizon-tls-certs\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.296899 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.315273 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chkvw\" (UniqueName: \"kubernetes.io/projected/4dd62da9-d481-4fc3-bbe6-108a54d98154-kube-api-access-chkvw\") pod \"horizon-6fd99d7d5b-sb96h\" (UID: \"4dd62da9-d481-4fc3-bbe6-108a54d98154\") " pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:25 crc kubenswrapper[4679]: I1202 10:38:25.390713 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:38:27 crc kubenswrapper[4679]: I1202 10:38:27.255693 4679 generic.go:334] "Generic (PLEG): container finished" podID="eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" containerID="81917a8d774a490d5fa6c3c95d1d994ec418a92049e6aeaafe5bf801b6794d0f" exitCode=0 Dec 02 10:38:27 crc kubenswrapper[4679]: I1202 10:38:27.255861 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rf2ql" event={"ID":"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3","Type":"ContainerDied","Data":"81917a8d774a490d5fa6c3c95d1d994ec418a92049e6aeaafe5bf801b6794d0f"} Dec 02 10:38:27 crc kubenswrapper[4679]: I1202 10:38:27.291691 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:38:27 crc kubenswrapper[4679]: I1202 10:38:27.353279 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sv4sp"] Dec 02 10:38:27 crc kubenswrapper[4679]: I1202 10:38:27.353852 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" containerID="cri-o://c53e9b861fa7f5dd3cf4c7650b06bf990c2be32822c8b90dfbc5af689f11ba1d" gracePeriod=10 Dec 02 10:38:28 crc kubenswrapper[4679]: I1202 10:38:28.271642 4679 generic.go:334] "Generic (PLEG): container finished" podID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerID="c53e9b861fa7f5dd3cf4c7650b06bf990c2be32822c8b90dfbc5af689f11ba1d" exitCode=0 Dec 02 10:38:28 crc kubenswrapper[4679]: I1202 10:38:28.271701 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" event={"ID":"11db932d-1c7f-4bdf-8400-47c562d2cff7","Type":"ContainerDied","Data":"c53e9b861fa7f5dd3cf4c7650b06bf990c2be32822c8b90dfbc5af689f11ba1d"} Dec 02 10:38:30 crc kubenswrapper[4679]: I1202 10:38:30.361136 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Dec 02 10:38:35 crc kubenswrapper[4679]: E1202 10:38:35.057523 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 02 10:38:35 crc kubenswrapper[4679]: E1202 10:38:35.058249 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n75hf7h9dh55fh68dh5c6hffh6hbh5c4h5h558hb6h58bh5f6h588hd5h684h5ch5ffh677h6dh9ch555hbbh657hb6h554h59hc6h659h68dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c29f7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-794d9b8d7f-pr9jf_openstack(6a750d67-054d-45d8-b1bb-30d89a141d72): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:35 crc kubenswrapper[4679]: E1202 10:38:35.061522 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-794d9b8d7f-pr9jf" podUID="6a750d67-054d-45d8-b1bb-30d89a141d72" Dec 02 10:38:35 crc kubenswrapper[4679]: E1202 10:38:35.340025 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 02 10:38:35 crc kubenswrapper[4679]: E1202 10:38:35.340199 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5h688h64ch687h9dh5c5h547h658h678h545h57fhcfh5b7h85h5bdh66fhb6h545h75h9bh64bh567h55fh54dh56bh67fh574h78h597h54bh696hcfq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hb9w6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(470284bf-e122-400f-9722-f8d4f23a75cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:35 crc kubenswrapper[4679]: I1202 10:38:35.360722 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Dec 02 10:38:37 crc kubenswrapper[4679]: E1202 10:38:37.001807 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 02 10:38:37 crc kubenswrapper[4679]: E1202 10:38:37.002422 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n666h5f9h67h5d5h578hfdh65ch66dh68dh684h577h5c5h5d9h56dhcbh57h544hc5hd6hfh74h7dhd9h648hd7h648h7fhd4h57ch556hbhc4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rzq2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-9b9b55977-kp8vd_openstack(b3e30d8f-75cb-474d-85b0-6656b987dae8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:37 crc kubenswrapper[4679]: E1202 10:38:37.004725 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-9b9b55977-kp8vd" podUID="b3e30d8f-75cb-474d-85b0-6656b987dae8" Dec 02 10:38:37 crc kubenswrapper[4679]: E1202 10:38:37.006648 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 02 10:38:37 crc kubenswrapper[4679]: E1202 10:38:37.006753 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n655hch9bh55h5f4hch678h7dh7dh698hc6h56bh5b4h684h688h8dh596h79hb5h649h64bh587h669h56h58h569h7fh5b9hc7h68hdfh546q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c2c95,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5447f744df-7bx4r_openstack(0b127b78-4f77-479e-ad9c-3abdf9bfdcc6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:37 crc kubenswrapper[4679]: E1202 10:38:37.008793 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5447f744df-7bx4r" podUID="0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" Dec 02 10:38:38 crc kubenswrapper[4679]: E1202 10:38:38.414441 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 02 10:38:38 crc kubenswrapper[4679]: E1202 10:38:38.414957 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qnxjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-9kjcb_openstack(055921c7-b108-40a9-a4da-2947ca27c2d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:38 crc kubenswrapper[4679]: E1202 10:38:38.416844 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-9kjcb" podUID="055921c7-b108-40a9-a4da-2947ca27c2d8" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.537328 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.546564 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-horizon-secret-key\") pod \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.546625 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2c95\" (UniqueName: \"kubernetes.io/projected/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-kube-api-access-c2c95\") pod \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.546659 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-config-data\") pod \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.546693 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-scripts\") pod \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.546718 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-logs\") pod \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\" (UID: \"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.547529 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-scripts" (OuterVolumeSpecName: "scripts") pod "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" (UID: "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.547597 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-config-data" (OuterVolumeSpecName: "config-data") pod "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" (UID: "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.547710 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-logs" (OuterVolumeSpecName: "logs") pod "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" (UID: "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.547740 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.547754 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.548618 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.557859 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" (UID: "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.584681 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-kube-api-access-c2c95" (OuterVolumeSpecName: "kube-api-access-c2c95") pod "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" (UID: "0b127b78-4f77-479e-ad9c-3abdf9bfdcc6"). InnerVolumeSpecName "kube-api-access-c2c95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.617133 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.624365 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648376 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3e30d8f-75cb-474d-85b0-6656b987dae8-logs\") pod \"b3e30d8f-75cb-474d-85b0-6656b987dae8\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648412 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-combined-ca-bundle\") pod \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648445 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-config-data\") pod \"b3e30d8f-75cb-474d-85b0-6656b987dae8\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648467 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-scripts\") pod \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648483 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c29f7\" (UniqueName: \"kubernetes.io/projected/6a750d67-054d-45d8-b1bb-30d89a141d72-kube-api-access-c29f7\") pod \"6a750d67-054d-45d8-b1bb-30d89a141d72\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648500 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-fernet-keys\") pod \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648526 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-224p2\" (UniqueName: \"kubernetes.io/projected/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-kube-api-access-224p2\") pod \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648551 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a750d67-054d-45d8-b1bb-30d89a141d72-logs\") pod \"6a750d67-054d-45d8-b1bb-30d89a141d72\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648568 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3e30d8f-75cb-474d-85b0-6656b987dae8-horizon-secret-key\") pod \"b3e30d8f-75cb-474d-85b0-6656b987dae8\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648598 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a750d67-054d-45d8-b1bb-30d89a141d72-horizon-secret-key\") pod \"6a750d67-054d-45d8-b1bb-30d89a141d72\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648625 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-config-data\") pod \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648660 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-scripts\") pod \"b3e30d8f-75cb-474d-85b0-6656b987dae8\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648676 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-config-data\") pod \"6a750d67-054d-45d8-b1bb-30d89a141d72\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648691 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzq2t\" (UniqueName: \"kubernetes.io/projected/b3e30d8f-75cb-474d-85b0-6656b987dae8-kube-api-access-rzq2t\") pod \"b3e30d8f-75cb-474d-85b0-6656b987dae8\" (UID: \"b3e30d8f-75cb-474d-85b0-6656b987dae8\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648686 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e30d8f-75cb-474d-85b0-6656b987dae8-logs" (OuterVolumeSpecName: "logs") pod "b3e30d8f-75cb-474d-85b0-6656b987dae8" (UID: "b3e30d8f-75cb-474d-85b0-6656b987dae8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648707 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-scripts\") pod \"6a750d67-054d-45d8-b1bb-30d89a141d72\" (UID: \"6a750d67-054d-45d8-b1bb-30d89a141d72\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.648770 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-credential-keys\") pod \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\" (UID: \"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3\") " Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.649509 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.649522 4679 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.649531 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3e30d8f-75cb-474d-85b0-6656b987dae8-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.649540 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2c95\" (UniqueName: \"kubernetes.io/projected/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6-kube-api-access-c2c95\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.649787 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a750d67-054d-45d8-b1bb-30d89a141d72-logs" (OuterVolumeSpecName: "logs") pod "6a750d67-054d-45d8-b1bb-30d89a141d72" (UID: "6a750d67-054d-45d8-b1bb-30d89a141d72"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.650629 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-scripts" (OuterVolumeSpecName: "scripts") pod "6a750d67-054d-45d8-b1bb-30d89a141d72" (UID: "6a750d67-054d-45d8-b1bb-30d89a141d72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.656009 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-scripts" (OuterVolumeSpecName: "scripts") pod "b3e30d8f-75cb-474d-85b0-6656b987dae8" (UID: "b3e30d8f-75cb-474d-85b0-6656b987dae8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.656184 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" (UID: "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.661007 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-scripts" (OuterVolumeSpecName: "scripts") pod "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" (UID: "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.661252 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a750d67-054d-45d8-b1bb-30d89a141d72-kube-api-access-c29f7" (OuterVolumeSpecName: "kube-api-access-c29f7") pod "6a750d67-054d-45d8-b1bb-30d89a141d72" (UID: "6a750d67-054d-45d8-b1bb-30d89a141d72"). InnerVolumeSpecName "kube-api-access-c29f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.661817 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-config-data" (OuterVolumeSpecName: "config-data") pod "b3e30d8f-75cb-474d-85b0-6656b987dae8" (UID: "b3e30d8f-75cb-474d-85b0-6656b987dae8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.661966 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-config-data" (OuterVolumeSpecName: "config-data") pod "6a750d67-054d-45d8-b1bb-30d89a141d72" (UID: "6a750d67-054d-45d8-b1bb-30d89a141d72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.662031 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" (UID: "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.662134 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e30d8f-75cb-474d-85b0-6656b987dae8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b3e30d8f-75cb-474d-85b0-6656b987dae8" (UID: "b3e30d8f-75cb-474d-85b0-6656b987dae8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.662851 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a750d67-054d-45d8-b1bb-30d89a141d72-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6a750d67-054d-45d8-b1bb-30d89a141d72" (UID: "6a750d67-054d-45d8-b1bb-30d89a141d72"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.669473 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e30d8f-75cb-474d-85b0-6656b987dae8-kube-api-access-rzq2t" (OuterVolumeSpecName: "kube-api-access-rzq2t") pod "b3e30d8f-75cb-474d-85b0-6656b987dae8" (UID: "b3e30d8f-75cb-474d-85b0-6656b987dae8"). InnerVolumeSpecName "kube-api-access-rzq2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.681480 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-kube-api-access-224p2" (OuterVolumeSpecName: "kube-api-access-224p2") pod "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" (UID: "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3"). InnerVolumeSpecName "kube-api-access-224p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.690982 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-config-data" (OuterVolumeSpecName: "config-data") pod "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" (UID: "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.712953 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" (UID: "eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751319 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751346 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751357 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzq2t\" (UniqueName: \"kubernetes.io/projected/b3e30d8f-75cb-474d-85b0-6656b987dae8-kube-api-access-rzq2t\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751367 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a750d67-054d-45d8-b1bb-30d89a141d72-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751375 4679 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751383 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751391 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3e30d8f-75cb-474d-85b0-6656b987dae8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751399 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751406 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c29f7\" (UniqueName: \"kubernetes.io/projected/6a750d67-054d-45d8-b1bb-30d89a141d72-kube-api-access-c29f7\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751414 4679 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751421 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-224p2\" (UniqueName: \"kubernetes.io/projected/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-kube-api-access-224p2\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751429 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a750d67-054d-45d8-b1bb-30d89a141d72-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751437 4679 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3e30d8f-75cb-474d-85b0-6656b987dae8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751444 4679 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6a750d67-054d-45d8-b1bb-30d89a141d72-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:38.751454 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:39 crc kubenswrapper[4679]: E1202 10:38:39.132239 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 02 10:38:39 crc kubenswrapper[4679]: E1202 10:38:39.132392 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pptfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-86lxg_openstack(f071f49e-c3e2-4919-a930-dd5fd654cf55): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:39 crc kubenswrapper[4679]: E1202 10:38:39.135761 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-86lxg" podUID="f071f49e-c3e2-4919-a930-dd5fd654cf55" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.375716 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rf2ql" event={"ID":"eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3","Type":"ContainerDied","Data":"8e12ea3a1bb7512e0d4f1e073ffc58e884d2c83a5e0978d3630bb44d1107dec3"} Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.375754 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e12ea3a1bb7512e0d4f1e073ffc58e884d2c83a5e0978d3630bb44d1107dec3" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.375776 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rf2ql" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.377892 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-794d9b8d7f-pr9jf" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.377936 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-794d9b8d7f-pr9jf" event={"ID":"6a750d67-054d-45d8-b1bb-30d89a141d72","Type":"ContainerDied","Data":"0521018c5eda668b1efddc8e3b13052a50d23d58e3e3de60b73e6962d063c77e"} Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.380360 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9b9b55977-kp8vd" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.380356 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9b9b55977-kp8vd" event={"ID":"b3e30d8f-75cb-474d-85b0-6656b987dae8","Type":"ContainerDied","Data":"bd11339b1cf56e2b1b79430cd0e9f966b664c361ab70a9a964cc0358a825beb3"} Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.381800 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5447f744df-7bx4r" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.381849 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5447f744df-7bx4r" event={"ID":"0b127b78-4f77-479e-ad9c-3abdf9bfdcc6","Type":"ContainerDied","Data":"cac18f4e70c3b8e8c2405e713d3809adea65a56e9065192c35edbeb7cd1f1c43"} Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.389708 4679 generic.go:334] "Generic (PLEG): container finished" podID="8d59898e-672f-4699-bf31-a964cd84913b" containerID="16c7e627e2b769fec6fded2a0edc40f16b7e1e272141289d48ed3bc7171d69c4" exitCode=0 Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.389769 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zs6tz" event={"ID":"8d59898e-672f-4699-bf31-a964cd84913b","Type":"ContainerDied","Data":"16c7e627e2b769fec6fded2a0edc40f16b7e1e272141289d48ed3bc7171d69c4"} Dec 02 10:38:39 crc kubenswrapper[4679]: E1202 10:38:39.393226 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-86lxg" podUID="f071f49e-c3e2-4919-a930-dd5fd654cf55" Dec 02 10:38:39 crc kubenswrapper[4679]: E1202 10:38:39.393389 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-9kjcb" podUID="055921c7-b108-40a9-a4da-2947ca27c2d8" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.451772 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9b9b55977-kp8vd"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.466175 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9b9b55977-kp8vd"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.504878 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-794d9b8d7f-pr9jf"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.520912 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-794d9b8d7f-pr9jf"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.561025 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5447f744df-7bx4r"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.568524 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5447f744df-7bx4r"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.710823 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rf2ql"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.717369 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rf2ql"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.797566 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9qtfk"] Dec 02 10:38:39 crc kubenswrapper[4679]: E1202 10:38:39.797991 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" containerName="keystone-bootstrap" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.798010 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" containerName="keystone-bootstrap" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.798223 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" containerName="keystone-bootstrap" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.798873 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.800719 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.800973 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.801107 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5x6pg" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.801397 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.801545 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.819407 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9qtfk"] Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.978115 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-fernet-keys\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.978181 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-combined-ca-bundle\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.978216 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-scripts\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.978240 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-config-data\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.978264 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-credential-keys\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:39 crc kubenswrapper[4679]: I1202 10:38:39.978300 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bttpn\" (UniqueName: \"kubernetes.io/projected/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-kube-api-access-bttpn\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.079399 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-combined-ca-bundle\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.079458 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-scripts\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.079482 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-config-data\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.079503 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-credential-keys\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.079535 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bttpn\" (UniqueName: \"kubernetes.io/projected/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-kube-api-access-bttpn\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.080226 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-fernet-keys\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.084389 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-credential-keys\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.084413 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-config-data\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.084470 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-fernet-keys\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.088582 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-scripts\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.088934 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-combined-ca-bundle\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.099018 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bttpn\" (UniqueName: \"kubernetes.io/projected/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-kube-api-access-bttpn\") pod \"keystone-bootstrap-9qtfk\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.140082 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.919407 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b127b78-4f77-479e-ad9c-3abdf9bfdcc6" path="/var/lib/kubelet/pods/0b127b78-4f77-479e-ad9c-3abdf9bfdcc6/volumes" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.920152 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a750d67-054d-45d8-b1bb-30d89a141d72" path="/var/lib/kubelet/pods/6a750d67-054d-45d8-b1bb-30d89a141d72/volumes" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.920595 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e30d8f-75cb-474d-85b0-6656b987dae8" path="/var/lib/kubelet/pods/b3e30d8f-75cb-474d-85b0-6656b987dae8/volumes" Dec 02 10:38:40 crc kubenswrapper[4679]: I1202 10:38:40.920961 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3" path="/var/lib/kubelet/pods/eb2a64dd-b5c8-4bf1-bd40-1a09e0ccf5e3/volumes" Dec 02 10:38:44 crc kubenswrapper[4679]: I1202 10:38:44.438526 4679 generic.go:334] "Generic (PLEG): container finished" podID="e1284622-943e-4f14-b59c-79eb084e8ea5" containerID="50450f1a8db17111a960666ee3f2b7e9231d0580c967b8b19c101e1179dc09a0" exitCode=0 Dec 02 10:38:44 crc kubenswrapper[4679]: I1202 10:38:44.438606 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-w58bz" event={"ID":"e1284622-943e-4f14-b59c-79eb084e8ea5","Type":"ContainerDied","Data":"50450f1a8db17111a960666ee3f2b7e9231d0580c967b8b19c101e1179dc09a0"} Dec 02 10:38:45 crc kubenswrapper[4679]: I1202 10:38:45.361564 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 02 10:38:45 crc kubenswrapper[4679]: I1202 10:38:45.361903 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:50 crc kubenswrapper[4679]: I1202 10:38:50.362828 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.646554 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.688730 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-config\") pod \"11db932d-1c7f-4bdf-8400-47c562d2cff7\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.688899 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-svc\") pod \"11db932d-1c7f-4bdf-8400-47c562d2cff7\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.688984 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-sb\") pod \"11db932d-1c7f-4bdf-8400-47c562d2cff7\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.689773 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-nb\") pod \"11db932d-1c7f-4bdf-8400-47c562d2cff7\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.689822 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwssg\" (UniqueName: \"kubernetes.io/projected/11db932d-1c7f-4bdf-8400-47c562d2cff7-kube-api-access-fwssg\") pod \"11db932d-1c7f-4bdf-8400-47c562d2cff7\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.689900 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-swift-storage-0\") pod \"11db932d-1c7f-4bdf-8400-47c562d2cff7\" (UID: \"11db932d-1c7f-4bdf-8400-47c562d2cff7\") " Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.696214 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11db932d-1c7f-4bdf-8400-47c562d2cff7-kube-api-access-fwssg" (OuterVolumeSpecName: "kube-api-access-fwssg") pod "11db932d-1c7f-4bdf-8400-47c562d2cff7" (UID: "11db932d-1c7f-4bdf-8400-47c562d2cff7"). InnerVolumeSpecName "kube-api-access-fwssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.738323 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11db932d-1c7f-4bdf-8400-47c562d2cff7" (UID: "11db932d-1c7f-4bdf-8400-47c562d2cff7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.746110 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11db932d-1c7f-4bdf-8400-47c562d2cff7" (UID: "11db932d-1c7f-4bdf-8400-47c562d2cff7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.757203 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11db932d-1c7f-4bdf-8400-47c562d2cff7" (UID: "11db932d-1c7f-4bdf-8400-47c562d2cff7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.766853 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-config" (OuterVolumeSpecName: "config") pod "11db932d-1c7f-4bdf-8400-47c562d2cff7" (UID: "11db932d-1c7f-4bdf-8400-47c562d2cff7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.775574 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "11db932d-1c7f-4bdf-8400-47c562d2cff7" (UID: "11db932d-1c7f-4bdf-8400-47c562d2cff7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.792409 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.792447 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.792458 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.792467 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.792479 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11db932d-1c7f-4bdf-8400-47c562d2cff7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.792488 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwssg\" (UniqueName: \"kubernetes.io/projected/11db932d-1c7f-4bdf-8400-47c562d2cff7-kube-api-access-fwssg\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:51 crc kubenswrapper[4679]: E1202 10:38:51.950475 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified" Dec 02 10:38:51 crc kubenswrapper[4679]: E1202 10:38:51.950639 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-notification-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5h688h64ch687h9dh5c5h547h658h678h545h57fhcfh5b7h85h5bdh66fhb6h545h75h9bh64bh567h55fh54dh56bh67fh574h78h597h54bh696hcfq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-notification-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hb9w6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/notificationhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(470284bf-e122-400f-9722-f8d4f23a75cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.973092 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zs6tz" Dec 02 10:38:51 crc kubenswrapper[4679]: I1202 10:38:51.980542 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.994806 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m69r2\" (UniqueName: \"kubernetes.io/projected/8d59898e-672f-4699-bf31-a964cd84913b-kube-api-access-m69r2\") pod \"8d59898e-672f-4699-bf31-a964cd84913b\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.994859 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdwxz\" (UniqueName: \"kubernetes.io/projected/e1284622-943e-4f14-b59c-79eb084e8ea5-kube-api-access-pdwxz\") pod \"e1284622-943e-4f14-b59c-79eb084e8ea5\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.995003 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-db-sync-config-data\") pod \"8d59898e-672f-4699-bf31-a964cd84913b\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.995031 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-config\") pod \"e1284622-943e-4f14-b59c-79eb084e8ea5\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.995104 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-config-data\") pod \"8d59898e-672f-4699-bf31-a964cd84913b\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.995173 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-combined-ca-bundle\") pod \"8d59898e-672f-4699-bf31-a964cd84913b\" (UID: \"8d59898e-672f-4699-bf31-a964cd84913b\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.995207 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-combined-ca-bundle\") pod \"e1284622-943e-4f14-b59c-79eb084e8ea5\" (UID: \"e1284622-943e-4f14-b59c-79eb084e8ea5\") " Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:51.998649 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8d59898e-672f-4699-bf31-a964cd84913b" (UID: "8d59898e-672f-4699-bf31-a964cd84913b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.000670 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1284622-943e-4f14-b59c-79eb084e8ea5-kube-api-access-pdwxz" (OuterVolumeSpecName: "kube-api-access-pdwxz") pod "e1284622-943e-4f14-b59c-79eb084e8ea5" (UID: "e1284622-943e-4f14-b59c-79eb084e8ea5"). InnerVolumeSpecName "kube-api-access-pdwxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.002130 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d59898e-672f-4699-bf31-a964cd84913b-kube-api-access-m69r2" (OuterVolumeSpecName: "kube-api-access-m69r2") pod "8d59898e-672f-4699-bf31-a964cd84913b" (UID: "8d59898e-672f-4699-bf31-a964cd84913b"). InnerVolumeSpecName "kube-api-access-m69r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.037181 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1284622-943e-4f14-b59c-79eb084e8ea5" (UID: "e1284622-943e-4f14-b59c-79eb084e8ea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.049832 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d59898e-672f-4699-bf31-a964cd84913b" (UID: "8d59898e-672f-4699-bf31-a964cd84913b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.068253 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-config-data" (OuterVolumeSpecName: "config-data") pod "8d59898e-672f-4699-bf31-a964cd84913b" (UID: "8d59898e-672f-4699-bf31-a964cd84913b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.079363 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-config" (OuterVolumeSpecName: "config") pod "e1284622-943e-4f14-b59c-79eb084e8ea5" (UID: "e1284622-943e-4f14-b59c-79eb084e8ea5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097511 4679 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097539 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097551 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097561 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d59898e-672f-4699-bf31-a964cd84913b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097572 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1284622-943e-4f14-b59c-79eb084e8ea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097582 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m69r2\" (UniqueName: \"kubernetes.io/projected/8d59898e-672f-4699-bf31-a964cd84913b-kube-api-access-m69r2\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.097595 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdwxz\" (UniqueName: \"kubernetes.io/projected/e1284622-943e-4f14-b59c-79eb084e8ea5-kube-api-access-pdwxz\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.409891 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fd99d7d5b-sb96h"] Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.515264 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" event={"ID":"11db932d-1c7f-4bdf-8400-47c562d2cff7","Type":"ContainerDied","Data":"b5d1940693da8a5a4204af0097908e9e02284e6c177c57a5161dad2c4cc59993"} Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.515324 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.515369 4679 scope.go:117] "RemoveContainer" containerID="c53e9b861fa7f5dd3cf4c7650b06bf990c2be32822c8b90dfbc5af689f11ba1d" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.529356 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zs6tz" event={"ID":"8d59898e-672f-4699-bf31-a964cd84913b","Type":"ContainerDied","Data":"79e47463117cf88a4f39a1c42956583d14fb2c06d9aa46a172fffb2481537a0c"} Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.529597 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79e47463117cf88a4f39a1c42956583d14fb2c06d9aa46a172fffb2481537a0c" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.529661 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zs6tz" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.536861 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-w58bz" event={"ID":"e1284622-943e-4f14-b59c-79eb084e8ea5","Type":"ContainerDied","Data":"1b1111d9d3d4e2aefa4c2a8224e76665ca258e1f9aedd5e79be07583962674ec"} Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.536927 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b1111d9d3d4e2aefa4c2a8224e76665ca258e1f9aedd5e79be07583962674ec" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.536989 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-w58bz" Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.632257 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sv4sp"] Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.640122 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sv4sp"] Dec 02 10:38:52 crc kubenswrapper[4679]: I1202 10:38:52.922123 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" path="/var/lib/kubelet/pods/11db932d-1c7f-4bdf-8400-47c562d2cff7/volumes" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.254330 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65965d6475-jj2h8"] Dec 02 10:38:53 crc kubenswrapper[4679]: E1202 10:38:53.254752 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1284622-943e-4f14-b59c-79eb084e8ea5" containerName="neutron-db-sync" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.254766 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1284622-943e-4f14-b59c-79eb084e8ea5" containerName="neutron-db-sync" Dec 02 10:38:53 crc kubenswrapper[4679]: E1202 10:38:53.254786 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="init" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.254794 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="init" Dec 02 10:38:53 crc kubenswrapper[4679]: E1202 10:38:53.254816 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.254824 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" Dec 02 10:38:53 crc kubenswrapper[4679]: E1202 10:38:53.254849 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d59898e-672f-4699-bf31-a964cd84913b" containerName="glance-db-sync" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.254858 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d59898e-672f-4699-bf31-a964cd84913b" containerName="glance-db-sync" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.255082 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1284622-943e-4f14-b59c-79eb084e8ea5" containerName="neutron-db-sync" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.255096 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d59898e-672f-4699-bf31-a964cd84913b" containerName="glance-db-sync" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.255111 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.256295 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.303670 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65965d6475-jj2h8"] Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.332859 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-svc\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.332918 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8bs4\" (UniqueName: \"kubernetes.io/projected/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-kube-api-access-l8bs4\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.332980 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-nb\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.333001 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-config\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.333030 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-swift-storage-0\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.333080 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-sb\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.435387 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-nb\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.435435 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-config\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.435462 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-swift-storage-0\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.435512 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-sb\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.435548 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-svc\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.435578 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8bs4\" (UniqueName: \"kubernetes.io/projected/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-kube-api-access-l8bs4\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.436745 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-sb\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.437227 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-svc\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.437527 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-nb\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.437767 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-config\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.438318 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-swift-storage-0\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.470779 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8bs4\" (UniqueName: \"kubernetes.io/projected/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-kube-api-access-l8bs4\") pod \"dnsmasq-dns-65965d6475-jj2h8\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.581842 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.637879 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-559ffdfb7b-75pb4"] Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.639263 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.643664 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.645913 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.646073 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9bb5d" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.646170 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.671006 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-559ffdfb7b-75pb4"] Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.739728 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65965d6475-jj2h8"] Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.747975 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-httpd-config\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.748200 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-ovndb-tls-certs\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.748269 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-config\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.748423 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-combined-ca-bundle\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.748566 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r62l\" (UniqueName: \"kubernetes.io/projected/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-kube-api-access-7r62l\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.777403 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-cltpj"] Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.778893 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.811856 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-cltpj"] Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.849948 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-combined-ca-bundle\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850004 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850077 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850104 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sfvv\" (UniqueName: \"kubernetes.io/projected/f1884c8a-197d-4631-a7bb-4fceb1f27649-kube-api-access-2sfvv\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850130 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r62l\" (UniqueName: \"kubernetes.io/projected/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-kube-api-access-7r62l\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850164 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-config\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850181 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850208 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850257 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-httpd-config\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850319 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-ovndb-tls-certs\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.850340 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-config\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.857056 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-combined-ca-bundle\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.857388 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-config\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.859879 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-ovndb-tls-certs\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.870173 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-httpd-config\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.873299 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r62l\" (UniqueName: \"kubernetes.io/projected/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-kube-api-access-7r62l\") pod \"neutron-559ffdfb7b-75pb4\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.951851 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sfvv\" (UniqueName: \"kubernetes.io/projected/f1884c8a-197d-4631-a7bb-4fceb1f27649-kube-api-access-2sfvv\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.953050 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-config\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.953122 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.953237 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.953420 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.953630 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.954848 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.954848 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.955795 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-config\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.956231 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.956712 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:53 crc kubenswrapper[4679]: I1202 10:38:53.979346 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sfvv\" (UniqueName: \"kubernetes.io/projected/f1884c8a-197d-4631-a7bb-4fceb1f27649-kube-api-access-2sfvv\") pod \"dnsmasq-dns-84b966f6c9-cltpj\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.012758 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.129350 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:54 crc kubenswrapper[4679]: E1202 10:38:54.357730 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 02 10:38:54 crc kubenswrapper[4679]: E1202 10:38:54.358413 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8ptzq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-bqb7p_openstack(048db06f-54db-4862-b8c8-b7e9cbb55095): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 10:38:54 crc kubenswrapper[4679]: E1202 10:38:54.359819 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-bqb7p" podUID="048db06f-54db-4862-b8c8-b7e9cbb55095" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.364024 4679 scope.go:117] "RemoveContainer" containerID="76c7d26940b8801e3d55aa1126f9cb0e60ef7aa2d32d3cfd779ee958041e86b9" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.472974 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.474604 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.477240 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.477619 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.477793 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-48wtd" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.479987 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.590205 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fd99d7d5b-sb96h" event={"ID":"4dd62da9-d481-4fc3-bbe6-108a54d98154","Type":"ContainerStarted","Data":"27dcfe4526e3f7836a9a8701b72f4af2ef1d93d37c6d5e4e88c2e9781eab985d"} Dec 02 10:38:54 crc kubenswrapper[4679]: E1202 10:38:54.593277 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-bqb7p" podUID="048db06f-54db-4862-b8c8-b7e9cbb55095" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672188 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-scripts\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672271 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672330 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672364 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-logs\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672402 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn2hp\" (UniqueName: \"kubernetes.io/projected/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-kube-api-access-pn2hp\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672420 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.672440 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-config-data\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.773575 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-scripts\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.773934 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.773974 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.774005 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-logs\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.774039 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn2hp\" (UniqueName: \"kubernetes.io/projected/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-kube-api-access-pn2hp\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.774057 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.774074 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-config-data\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.774922 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-logs\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.775371 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.776014 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.793580 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn2hp\" (UniqueName: \"kubernetes.io/projected/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-kube-api-access-pn2hp\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.794349 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-config-data\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.796522 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-scripts\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.813336 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:54 crc kubenswrapper[4679]: I1202 10:38:54.839260 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " pod="openstack/glance-default-external-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.025498 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9qtfk"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.074811 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.076711 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.080733 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.097968 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bd7d57448-84dtd"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.100329 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.121437 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.149859 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-559ffdfb7b-75pb4"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.202477 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.203506 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-logs\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.203687 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.203830 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.204362 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.204702 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-cltpj"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.205130 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzbpx\" (UniqueName: \"kubernetes.io/projected/3d63d49d-2614-4329-abde-118b126b9a0d-kube-api-access-xzbpx\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.205188 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: W1202 10:38:55.241444 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d11f0a0_986a_4012_bea2_a6c5b325cb7a.slice/crio-b3cfde8b41fe8c0ebd88b94cae96b301b8f7b41d0b3b24851e19004ec57e8bdc WatchSource:0}: Error finding container b3cfde8b41fe8c0ebd88b94cae96b301b8f7b41d0b3b24851e19004ec57e8bdc: Status 404 returned error can't find the container with id b3cfde8b41fe8c0ebd88b94cae96b301b8f7b41d0b3b24851e19004ec57e8bdc Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319177 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzbpx\" (UniqueName: \"kubernetes.io/projected/3d63d49d-2614-4329-abde-118b126b9a0d-kube-api-access-xzbpx\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319470 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319586 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319676 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-logs\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319734 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319785 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319804 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.319989 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.320291 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-logs\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.323635 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.333853 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.340811 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.344541 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65965d6475-jj2h8"] Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.372103 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzbpx\" (UniqueName: \"kubernetes.io/projected/3d63d49d-2614-4329-abde-118b126b9a0d-kube-api-access-xzbpx\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.372244 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sv4sp" podUID="11db932d-1c7f-4bdf-8400-47c562d2cff7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.375950 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.399743 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.538976 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.664610 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-86lxg" event={"ID":"f071f49e-c3e2-4919-a930-dd5fd654cf55","Type":"ContainerStarted","Data":"02285a941c9dc5583f689ea9856449177dc4724a9eda53ccf9257492f7eedcce"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.685202 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bd7d57448-84dtd" event={"ID":"ded5eaca-d45b-46bf-a770-50d909f0003e","Type":"ContainerStarted","Data":"9ddb76da0ced0815c162346f9bf7c3e84a84e82dc2aa9d29019c4c5218bedada"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.717283 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-86lxg" podStartSLOduration=3.320213837 podStartE2EDuration="39.717264877s" podCreationTimestamp="2025-12-02 10:38:16 +0000 UTC" firstStartedPulling="2025-12-02 10:38:18.111648481 +0000 UTC m=+1151.441787341" lastFinishedPulling="2025-12-02 10:38:54.508699521 +0000 UTC m=+1187.838838381" observedRunningTime="2025-12-02 10:38:55.710855692 +0000 UTC m=+1189.040994552" watchObservedRunningTime="2025-12-02 10:38:55.717264877 +0000 UTC m=+1189.047403737" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.724134 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-559ffdfb7b-75pb4" event={"ID":"3d11f0a0-986a-4012-bea2-a6c5b325cb7a","Type":"ContainerStarted","Data":"b3cfde8b41fe8c0ebd88b94cae96b301b8f7b41d0b3b24851e19004ec57e8bdc"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.733166 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fd99d7d5b-sb96h" event={"ID":"4dd62da9-d481-4fc3-bbe6-108a54d98154","Type":"ContainerStarted","Data":"7a39588e765c1cdc2c97b35054fbd0e980434594d724fb52f9786e661232a488"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.734897 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" event={"ID":"fb9b7374-1a57-46a3-aa55-313e02b7d4e3","Type":"ContainerStarted","Data":"e1349752cff8d461cc00ae16eabb17e0f7831260f7cef1e3c8740380afba508a"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.737530 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qtfk" event={"ID":"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b","Type":"ContainerStarted","Data":"4807083e9b2f93e0679d1db73827ffaa4974fc6cc8c70e41542529d1e28db2f8"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.737566 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qtfk" event={"ID":"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b","Type":"ContainerStarted","Data":"c7a5f9ed3a25cb79040d034f04a86b18dd17953bc1f25331620cb4bc51b85287"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.761455 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" event={"ID":"f1884c8a-197d-4631-a7bb-4fceb1f27649","Type":"ContainerStarted","Data":"7a1beea65d596f38ba363116a50fce95752e03ca4845ffab5962eed607c13232"} Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.780784 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9qtfk" podStartSLOduration=16.780768929 podStartE2EDuration="16.780768929s" podCreationTimestamp="2025-12-02 10:38:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:55.780736018 +0000 UTC m=+1189.110874878" watchObservedRunningTime="2025-12-02 10:38:55.780768929 +0000 UTC m=+1189.110907789" Dec 02 10:38:55 crc kubenswrapper[4679]: I1202 10:38:55.891182 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.382491 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.827174 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3d63d49d-2614-4329-abde-118b126b9a0d","Type":"ContainerStarted","Data":"ab0f364315ee7a6be3b1eb37dcdc12a92412b89ac227b3f095cb544cb5602e1d"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.843292 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6","Type":"ContainerStarted","Data":"63cfe4f96f0e292bb0deab3c0efe2b97548581291198c98c12336f77da681531"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.850949 4679 generic.go:334] "Generic (PLEG): container finished" podID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerID="335c4337cafedbf9d06c3114c91da5fa28cff05b64e159adc246b627b4961969" exitCode=0 Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.851043 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" event={"ID":"f1884c8a-197d-4631-a7bb-4fceb1f27649","Type":"ContainerDied","Data":"335c4337cafedbf9d06c3114c91da5fa28cff05b64e159adc246b627b4961969"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.876804 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9kjcb" event={"ID":"055921c7-b108-40a9-a4da-2947ca27c2d8","Type":"ContainerStarted","Data":"3dfec170a3ca144e37522da5e343af3fd966f5d6066ed9357992498cfa914b8d"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.884794 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.908727 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bd7d57448-84dtd" event={"ID":"ded5eaca-d45b-46bf-a770-50d909f0003e","Type":"ContainerStarted","Data":"d87f86cc90eb58eaa5a8aa14c2723787b723ae1ed4cfb00f1c88c0c427250f59"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.976921 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.977363 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-559ffdfb7b-75pb4" event={"ID":"3d11f0a0-986a-4012-bea2-a6c5b325cb7a","Type":"ContainerStarted","Data":"7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.977386 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-559ffdfb7b-75pb4" event={"ID":"3d11f0a0-986a-4012-bea2-a6c5b325cb7a","Type":"ContainerStarted","Data":"808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.977397 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fd99d7d5b-sb96h" event={"ID":"4dd62da9-d481-4fc3-bbe6-108a54d98154","Type":"ContainerStarted","Data":"5ef5632a658f02fe1b24a06b746ffd704b4814a0b2a6f7fe8f9a1a7c16b7a1c0"} Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.986827 4679 generic.go:334] "Generic (PLEG): container finished" podID="fb9b7374-1a57-46a3-aa55-313e02b7d4e3" containerID="7f42f7f22091b7f27772d541872b9300605a8a6390ccb951123573ef74b38b51" exitCode=0 Dec 02 10:38:56 crc kubenswrapper[4679]: I1202 10:38:56.987133 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" event={"ID":"fb9b7374-1a57-46a3-aa55-313e02b7d4e3","Type":"ContainerDied","Data":"7f42f7f22091b7f27772d541872b9300605a8a6390ccb951123573ef74b38b51"} Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.011674 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-9kjcb" podStartSLOduration=3.300649919 podStartE2EDuration="41.011650007s" podCreationTimestamp="2025-12-02 10:38:16 +0000 UTC" firstStartedPulling="2025-12-02 10:38:17.982582296 +0000 UTC m=+1151.312721156" lastFinishedPulling="2025-12-02 10:38:55.693582384 +0000 UTC m=+1189.023721244" observedRunningTime="2025-12-02 10:38:56.966895996 +0000 UTC m=+1190.297034856" watchObservedRunningTime="2025-12-02 10:38:57.011650007 +0000 UTC m=+1190.341788867" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.021259 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6bd7d57448-84dtd" podStartSLOduration=33.021240494 podStartE2EDuration="33.021240494s" podCreationTimestamp="2025-12-02 10:38:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:57.001611868 +0000 UTC m=+1190.331750718" watchObservedRunningTime="2025-12-02 10:38:57.021240494 +0000 UTC m=+1190.351379354" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.026978 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.035327 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-559ffdfb7b-75pb4" podStartSLOduration=4.03531189 podStartE2EDuration="4.03531189s" podCreationTimestamp="2025-12-02 10:38:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:57.028266247 +0000 UTC m=+1190.358405107" watchObservedRunningTime="2025-12-02 10:38:57.03531189 +0000 UTC m=+1190.365450740" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.185080 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6fd99d7d5b-sb96h" podStartSLOduration=32.618590798 podStartE2EDuration="33.18505749s" podCreationTimestamp="2025-12-02 10:38:24 +0000 UTC" firstStartedPulling="2025-12-02 10:38:54.332729795 +0000 UTC m=+1187.662868655" lastFinishedPulling="2025-12-02 10:38:54.899196487 +0000 UTC m=+1188.229335347" observedRunningTime="2025-12-02 10:38:57.114648479 +0000 UTC m=+1190.444787339" watchObservedRunningTime="2025-12-02 10:38:57.18505749 +0000 UTC m=+1190.515196350" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.460885 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.541032 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-config\") pod \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.541093 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-sb\") pod \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.541138 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8bs4\" (UniqueName: \"kubernetes.io/projected/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-kube-api-access-l8bs4\") pod \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.541237 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-svc\") pod \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.541266 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-nb\") pod \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.566772 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-kube-api-access-l8bs4" (OuterVolumeSpecName: "kube-api-access-l8bs4") pod "fb9b7374-1a57-46a3-aa55-313e02b7d4e3" (UID: "fb9b7374-1a57-46a3-aa55-313e02b7d4e3"). InnerVolumeSpecName "kube-api-access-l8bs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.582947 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-config" (OuterVolumeSpecName: "config") pod "fb9b7374-1a57-46a3-aa55-313e02b7d4e3" (UID: "fb9b7374-1a57-46a3-aa55-313e02b7d4e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.583740 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fb9b7374-1a57-46a3-aa55-313e02b7d4e3" (UID: "fb9b7374-1a57-46a3-aa55-313e02b7d4e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.597137 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fb9b7374-1a57-46a3-aa55-313e02b7d4e3" (UID: "fb9b7374-1a57-46a3-aa55-313e02b7d4e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.599655 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fb9b7374-1a57-46a3-aa55-313e02b7d4e3" (UID: "fb9b7374-1a57-46a3-aa55-313e02b7d4e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.642394 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-swift-storage-0\") pod \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\" (UID: \"fb9b7374-1a57-46a3-aa55-313e02b7d4e3\") " Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.642911 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.642928 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8bs4\" (UniqueName: \"kubernetes.io/projected/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-kube-api-access-l8bs4\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.642938 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.642950 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.642959 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.677984 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fb9b7374-1a57-46a3-aa55-313e02b7d4e3" (UID: "fb9b7374-1a57-46a3-aa55-313e02b7d4e3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:38:57 crc kubenswrapper[4679]: I1202 10:38:57.748341 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9b7374-1a57-46a3-aa55-313e02b7d4e3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.015688 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" event={"ID":"fb9b7374-1a57-46a3-aa55-313e02b7d4e3","Type":"ContainerDied","Data":"e1349752cff8d461cc00ae16eabb17e0f7831260f7cef1e3c8740380afba508a"} Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.015748 4679 scope.go:117] "RemoveContainer" containerID="7f42f7f22091b7f27772d541872b9300605a8a6390ccb951123573ef74b38b51" Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.015895 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65965d6475-jj2h8" Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.026433 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3d63d49d-2614-4329-abde-118b126b9a0d","Type":"ContainerStarted","Data":"fbb27f2ae7c06d8c39442eafd4cc9b95d377a64f8a448fb9e9ad49970b9d4a05"} Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.030890 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6","Type":"ContainerStarted","Data":"d224426f9c63cc6361661cb816686582918028166e0fe3ecb9fd9bc9723e1441"} Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.036377 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" event={"ID":"f1884c8a-197d-4631-a7bb-4fceb1f27649","Type":"ContainerStarted","Data":"00ff65400adb4603bece89da8eaf7863189ced756ba99423db2081b70e9255c4"} Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.038745 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.063211 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bd7d57448-84dtd" event={"ID":"ded5eaca-d45b-46bf-a770-50d909f0003e","Type":"ContainerStarted","Data":"dbbabf0a29747ebfd44364b9b3f7b60a84d413aee3ffe565fee42f380f0ae98f"} Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.094698 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" podStartSLOduration=5.094669391 podStartE2EDuration="5.094669391s" podCreationTimestamp="2025-12-02 10:38:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:58.074968662 +0000 UTC m=+1191.405107522" watchObservedRunningTime="2025-12-02 10:38:58.094669391 +0000 UTC m=+1191.424808251" Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.193721 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65965d6475-jj2h8"] Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.221349 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65965d6475-jj2h8"] Dec 02 10:38:58 crc kubenswrapper[4679]: I1202 10:38:58.949078 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb9b7374-1a57-46a3-aa55-313e02b7d4e3" path="/var/lib/kubelet/pods/fb9b7374-1a57-46a3-aa55-313e02b7d4e3/volumes" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.080147 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6","Type":"ContainerStarted","Data":"7fba9979b37e03c958764d8b9abf6f4371e41629ba605915b95e8fb5e437dde8"} Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.080326 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-log" containerID="cri-o://d224426f9c63cc6361661cb816686582918028166e0fe3ecb9fd9bc9723e1441" gracePeriod=30 Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.080954 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-httpd" containerID="cri-o://7fba9979b37e03c958764d8b9abf6f4371e41629ba605915b95e8fb5e437dde8" gracePeriod=30 Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.098713 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-log" containerID="cri-o://fbb27f2ae7c06d8c39442eafd4cc9b95d377a64f8a448fb9e9ad49970b9d4a05" gracePeriod=30 Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.098877 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-httpd" containerID="cri-o://16a16785b1a0ad37fb80eaba358e35f064a1f75277a951eab1e2d28a63824d75" gracePeriod=30 Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.098978 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3d63d49d-2614-4329-abde-118b126b9a0d","Type":"ContainerStarted","Data":"16a16785b1a0ad37fb80eaba358e35f064a1f75277a951eab1e2d28a63824d75"} Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.130340 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.130317978 podStartE2EDuration="6.130317978s" podCreationTimestamp="2025-12-02 10:38:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:59.120798523 +0000 UTC m=+1192.450937393" watchObservedRunningTime="2025-12-02 10:38:59.130317978 +0000 UTC m=+1192.460456838" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.166008 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.165986737 podStartE2EDuration="5.165986737s" podCreationTimestamp="2025-12-02 10:38:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:38:59.165612016 +0000 UTC m=+1192.495750886" watchObservedRunningTime="2025-12-02 10:38:59.165986737 +0000 UTC m=+1192.496125597" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.769975 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6dbdc9c48f-dtqv6"] Dec 02 10:38:59 crc kubenswrapper[4679]: E1202 10:38:59.770449 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9b7374-1a57-46a3-aa55-313e02b7d4e3" containerName="init" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.770469 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9b7374-1a57-46a3-aa55-313e02b7d4e3" containerName="init" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.770679 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9b7374-1a57-46a3-aa55-313e02b7d4e3" containerName="init" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.771679 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.774460 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.786495 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.793717 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dbdc9c48f-dtqv6"] Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911466 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-internal-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911582 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-httpd-config\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911602 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-config\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911635 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-combined-ca-bundle\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911652 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-ovndb-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911700 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftdln\" (UniqueName: \"kubernetes.io/projected/eecb1bea-c78b-4888-b9a8-e868ab82a9db-kube-api-access-ftdln\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:38:59 crc kubenswrapper[4679]: I1202 10:38:59.911726 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-public-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.012597 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-public-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.012854 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-internal-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.012963 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-httpd-config\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.012979 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-config\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.013013 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-combined-ca-bundle\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.013028 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-ovndb-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.013083 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftdln\" (UniqueName: \"kubernetes.io/projected/eecb1bea-c78b-4888-b9a8-e868ab82a9db-kube-api-access-ftdln\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.022127 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-internal-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.027330 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-combined-ca-bundle\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.029643 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-httpd-config\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.031904 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-config\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.037928 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftdln\" (UniqueName: \"kubernetes.io/projected/eecb1bea-c78b-4888-b9a8-e868ab82a9db-kube-api-access-ftdln\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.042621 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-ovndb-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.053881 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eecb1bea-c78b-4888-b9a8-e868ab82a9db-public-tls-certs\") pod \"neutron-6dbdc9c48f-dtqv6\" (UID: \"eecb1bea-c78b-4888-b9a8-e868ab82a9db\") " pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.108406 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.118999 4679 generic.go:334] "Generic (PLEG): container finished" podID="3d63d49d-2614-4329-abde-118b126b9a0d" containerID="16a16785b1a0ad37fb80eaba358e35f064a1f75277a951eab1e2d28a63824d75" exitCode=0 Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.119035 4679 generic.go:334] "Generic (PLEG): container finished" podID="3d63d49d-2614-4329-abde-118b126b9a0d" containerID="fbb27f2ae7c06d8c39442eafd4cc9b95d377a64f8a448fb9e9ad49970b9d4a05" exitCode=143 Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.119103 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3d63d49d-2614-4329-abde-118b126b9a0d","Type":"ContainerDied","Data":"16a16785b1a0ad37fb80eaba358e35f064a1f75277a951eab1e2d28a63824d75"} Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.119133 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3d63d49d-2614-4329-abde-118b126b9a0d","Type":"ContainerDied","Data":"fbb27f2ae7c06d8c39442eafd4cc9b95d377a64f8a448fb9e9ad49970b9d4a05"} Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.122383 4679 generic.go:334] "Generic (PLEG): container finished" podID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerID="7fba9979b37e03c958764d8b9abf6f4371e41629ba605915b95e8fb5e437dde8" exitCode=0 Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.122440 4679 generic.go:334] "Generic (PLEG): container finished" podID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerID="d224426f9c63cc6361661cb816686582918028166e0fe3ecb9fd9bc9723e1441" exitCode=143 Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.122520 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6","Type":"ContainerDied","Data":"7fba9979b37e03c958764d8b9abf6f4371e41629ba605915b95e8fb5e437dde8"} Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.122543 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6","Type":"ContainerDied","Data":"d224426f9c63cc6361661cb816686582918028166e0fe3ecb9fd9bc9723e1441"} Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.128742 4679 generic.go:334] "Generic (PLEG): container finished" podID="f071f49e-c3e2-4919-a930-dd5fd654cf55" containerID="02285a941c9dc5583f689ea9856449177dc4724a9eda53ccf9257492f7eedcce" exitCode=0 Dec 02 10:39:00 crc kubenswrapper[4679]: I1202 10:39:00.128818 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-86lxg" event={"ID":"f071f49e-c3e2-4919-a930-dd5fd654cf55","Type":"ContainerDied","Data":"02285a941c9dc5583f689ea9856449177dc4724a9eda53ccf9257492f7eedcce"} Dec 02 10:39:01 crc kubenswrapper[4679]: I1202 10:39:01.154811 4679 generic.go:334] "Generic (PLEG): container finished" podID="6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" containerID="4807083e9b2f93e0679d1db73827ffaa4974fc6cc8c70e41542529d1e28db2f8" exitCode=0 Dec 02 10:39:01 crc kubenswrapper[4679]: I1202 10:39:01.155088 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qtfk" event={"ID":"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b","Type":"ContainerDied","Data":"4807083e9b2f93e0679d1db73827ffaa4974fc6cc8c70e41542529d1e28db2f8"} Dec 02 10:39:01 crc kubenswrapper[4679]: I1202 10:39:01.158873 4679 generic.go:334] "Generic (PLEG): container finished" podID="055921c7-b108-40a9-a4da-2947ca27c2d8" containerID="3dfec170a3ca144e37522da5e343af3fd966f5d6066ed9357992498cfa914b8d" exitCode=0 Dec 02 10:39:01 crc kubenswrapper[4679]: I1202 10:39:01.159084 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9kjcb" event={"ID":"055921c7-b108-40a9-a4da-2947ca27c2d8","Type":"ContainerDied","Data":"3dfec170a3ca144e37522da5e343af3fd966f5d6066ed9357992498cfa914b8d"} Dec 02 10:39:01 crc kubenswrapper[4679]: I1202 10:39:01.952130 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.059870 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-httpd-run\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060024 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060065 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-config-data\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060105 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-scripts\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060167 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn2hp\" (UniqueName: \"kubernetes.io/projected/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-kube-api-access-pn2hp\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060183 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-logs\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060218 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-combined-ca-bundle\") pod \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\" (UID: \"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6\") " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060552 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-logs" (OuterVolumeSpecName: "logs") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.060585 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.061218 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.061240 4679 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.078646 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-kube-api-access-pn2hp" (OuterVolumeSpecName: "kube-api-access-pn2hp") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "kube-api-access-pn2hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.080454 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.081840 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-scripts" (OuterVolumeSpecName: "scripts") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.088117 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.106010 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-config-data" (OuterVolumeSpecName: "config-data") pod "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" (UID: "0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.163465 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.163504 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.163513 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.163523 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn2hp\" (UniqueName: \"kubernetes.io/projected/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-kube-api-access-pn2hp\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.163534 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.172347 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.174218 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6","Type":"ContainerDied","Data":"63cfe4f96f0e292bb0deab3c0efe2b97548581291198c98c12336f77da681531"} Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.174406 4679 scope.go:117] "RemoveContainer" containerID="7fba9979b37e03c958764d8b9abf6f4371e41629ba605915b95e8fb5e437dde8" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.180968 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.216054 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.225288 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.262667 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:02 crc kubenswrapper[4679]: E1202 10:39:02.263542 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-log" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.263570 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-log" Dec 02 10:39:02 crc kubenswrapper[4679]: E1202 10:39:02.263591 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-httpd" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.263601 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-httpd" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.265219 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.265525 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-log" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.265564 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" containerName="glance-httpd" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.267236 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.300740 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.300999 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.304184 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468107 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-logs\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468508 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468602 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468637 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468655 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468739 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbvm6\" (UniqueName: \"kubernetes.io/projected/c5799263-7d7f-4195-b6b2-70d9315e3a3f-kube-api-access-sbvm6\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468756 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.468826 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570438 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570503 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-logs\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570524 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570575 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570593 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570610 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570664 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbvm6\" (UniqueName: \"kubernetes.io/projected/c5799263-7d7f-4195-b6b2-70d9315e3a3f-kube-api-access-sbvm6\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.570679 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.572858 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.574941 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.577571 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-logs\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.580951 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.581098 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.582014 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.584035 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.604155 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbvm6\" (UniqueName: \"kubernetes.io/projected/c5799263-7d7f-4195-b6b2-70d9315e3a3f-kube-api-access-sbvm6\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.683554 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.914837 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:02 crc kubenswrapper[4679]: I1202 10:39:02.920691 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6" path="/var/lib/kubelet/pods/0bec9552-dfaa-4fe3-b3c6-82fe6fea40a6/volumes" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.131455 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.206502 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jtqff"] Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.206784 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" containerName="dnsmasq-dns" containerID="cri-o://5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9" gracePeriod=10 Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.356751 4679 scope.go:117] "RemoveContainer" containerID="d224426f9c63cc6361661cb816686582918028166e0fe3ecb9fd9bc9723e1441" Dec 02 10:39:04 crc kubenswrapper[4679]: E1202 10:39:04.474404 4679 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78457c3d_d7fe_413c_aca4_63a57deec717.slice/crio-5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78457c3d_d7fe_413c_aca4_63a57deec717.slice/crio-conmon-5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9.scope\": RecentStats: unable to find data in memory cache]" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.653883 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9kjcb" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.678070 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-86lxg" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.693826 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.717843 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.736627 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzbpx\" (UniqueName: \"kubernetes.io/projected/3d63d49d-2614-4329-abde-118b126b9a0d-kube-api-access-xzbpx\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741476 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/055921c7-b108-40a9-a4da-2947ca27c2d8-logs\") pod \"055921c7-b108-40a9-a4da-2947ca27c2d8\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741539 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-config-data\") pod \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741568 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-combined-ca-bundle\") pod \"f071f49e-c3e2-4919-a930-dd5fd654cf55\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741604 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-combined-ca-bundle\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741640 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-httpd-run\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741665 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pptfr\" (UniqueName: \"kubernetes.io/projected/f071f49e-c3e2-4919-a930-dd5fd654cf55-kube-api-access-pptfr\") pod \"f071f49e-c3e2-4919-a930-dd5fd654cf55\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741691 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-credential-keys\") pod \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741740 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-combined-ca-bundle\") pod \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741757 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-scripts\") pod \"055921c7-b108-40a9-a4da-2947ca27c2d8\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741782 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-db-sync-config-data\") pod \"f071f49e-c3e2-4919-a930-dd5fd654cf55\" (UID: \"f071f49e-c3e2-4919-a930-dd5fd654cf55\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.741809 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bttpn\" (UniqueName: \"kubernetes.io/projected/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-kube-api-access-bttpn\") pod \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.743688 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.743953 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/055921c7-b108-40a9-a4da-2947ca27c2d8-logs" (OuterVolumeSpecName: "logs") pod "055921c7-b108-40a9-a4da-2947ca27c2d8" (UID: "055921c7-b108-40a9-a4da-2947ca27c2d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.746552 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d63d49d-2614-4329-abde-118b126b9a0d-kube-api-access-xzbpx" (OuterVolumeSpecName: "kube-api-access-xzbpx") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "kube-api-access-xzbpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.754474 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-scripts" (OuterVolumeSpecName: "scripts") pod "055921c7-b108-40a9-a4da-2947ca27c2d8" (UID: "055921c7-b108-40a9-a4da-2947ca27c2d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.757694 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-kube-api-access-bttpn" (OuterVolumeSpecName: "kube-api-access-bttpn") pod "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" (UID: "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b"). InnerVolumeSpecName "kube-api-access-bttpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.760476 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f071f49e-c3e2-4919-a930-dd5fd654cf55" (UID: "f071f49e-c3e2-4919-a930-dd5fd654cf55"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.772732 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f071f49e-c3e2-4919-a930-dd5fd654cf55-kube-api-access-pptfr" (OuterVolumeSpecName: "kube-api-access-pptfr") pod "f071f49e-c3e2-4919-a930-dd5fd654cf55" (UID: "f071f49e-c3e2-4919-a930-dd5fd654cf55"). InnerVolumeSpecName "kube-api-access-pptfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.776178 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" (UID: "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.784256 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.802705 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-config-data" (OuterVolumeSpecName: "config-data") pod "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" (UID: "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.805783 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" (UID: "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.820255 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f071f49e-c3e2-4919-a930-dd5fd654cf55" (UID: "f071f49e-c3e2-4919-a930-dd5fd654cf55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.834297 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.844770 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-combined-ca-bundle\") pod \"055921c7-b108-40a9-a4da-2947ca27c2d8\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.844976 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845016 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-fernet-keys\") pod \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845032 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-scripts\") pod \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\" (UID: \"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845077 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-config-data\") pod \"055921c7-b108-40a9-a4da-2947ca27c2d8\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845106 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-config-data\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845129 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnxjd\" (UniqueName: \"kubernetes.io/projected/055921c7-b108-40a9-a4da-2947ca27c2d8-kube-api-access-qnxjd\") pod \"055921c7-b108-40a9-a4da-2947ca27c2d8\" (UID: \"055921c7-b108-40a9-a4da-2947ca27c2d8\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845151 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-scripts\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845194 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-logs\") pod \"3d63d49d-2614-4329-abde-118b126b9a0d\" (UID: \"3d63d49d-2614-4329-abde-118b126b9a0d\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845447 4679 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845464 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bttpn\" (UniqueName: \"kubernetes.io/projected/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-kube-api-access-bttpn\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845474 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzbpx\" (UniqueName: \"kubernetes.io/projected/3d63d49d-2614-4329-abde-118b126b9a0d-kube-api-access-xzbpx\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845482 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/055921c7-b108-40a9-a4da-2947ca27c2d8-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845492 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845500 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f071f49e-c3e2-4919-a930-dd5fd654cf55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845508 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845516 4679 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845524 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pptfr\" (UniqueName: \"kubernetes.io/projected/f071f49e-c3e2-4919-a930-dd5fd654cf55-kube-api-access-pptfr\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845532 4679 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845539 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845547 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.845941 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-logs" (OuterVolumeSpecName: "logs") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.850044 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-scripts" (OuterVolumeSpecName: "scripts") pod "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" (UID: "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.851447 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/055921c7-b108-40a9-a4da-2947ca27c2d8-kube-api-access-qnxjd" (OuterVolumeSpecName: "kube-api-access-qnxjd") pod "055921c7-b108-40a9-a4da-2947ca27c2d8" (UID: "055921c7-b108-40a9-a4da-2947ca27c2d8"). InnerVolumeSpecName "kube-api-access-qnxjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.855548 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.856685 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-scripts" (OuterVolumeSpecName: "scripts") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.866275 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" (UID: "6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.901780 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-config-data" (OuterVolumeSpecName: "config-data") pod "055921c7-b108-40a9-a4da-2947ca27c2d8" (UID: "055921c7-b108-40a9-a4da-2947ca27c2d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.910981 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "055921c7-b108-40a9-a4da-2947ca27c2d8" (UID: "055921c7-b108-40a9-a4da-2947ca27c2d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.950148 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzxwl\" (UniqueName: \"kubernetes.io/projected/78457c3d-d7fe-413c-aca4-63a57deec717-kube-api-access-lzxwl\") pod \"78457c3d-d7fe-413c-aca4-63a57deec717\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.952603 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-swift-storage-0\") pod \"78457c3d-d7fe-413c-aca4-63a57deec717\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.952724 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-nb\") pod \"78457c3d-d7fe-413c-aca4-63a57deec717\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.952970 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-sb\") pod \"78457c3d-d7fe-413c-aca4-63a57deec717\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.953015 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-svc\") pod \"78457c3d-d7fe-413c-aca4-63a57deec717\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.953065 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-config\") pod \"78457c3d-d7fe-413c-aca4-63a57deec717\" (UID: \"78457c3d-d7fe-413c-aca4-63a57deec717\") " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954498 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78457c3d-d7fe-413c-aca4-63a57deec717-kube-api-access-lzxwl" (OuterVolumeSpecName: "kube-api-access-lzxwl") pod "78457c3d-d7fe-413c-aca4-63a57deec717" (UID: "78457c3d-d7fe-413c-aca4-63a57deec717"). InnerVolumeSpecName "kube-api-access-lzxwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954712 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954750 4679 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954781 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954797 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954808 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnxjd\" (UniqueName: \"kubernetes.io/projected/055921c7-b108-40a9-a4da-2947ca27c2d8-kube-api-access-qnxjd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954818 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954827 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d63d49d-2614-4329-abde-118b126b9a0d-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.954863 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/055921c7-b108-40a9-a4da-2947ca27c2d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.969250 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-config-data" (OuterVolumeSpecName: "config-data") pod "3d63d49d-2614-4329-abde-118b126b9a0d" (UID: "3d63d49d-2614-4329-abde-118b126b9a0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:04 crc kubenswrapper[4679]: I1202 10:39:04.983340 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.008077 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "78457c3d-d7fe-413c-aca4-63a57deec717" (UID: "78457c3d-d7fe-413c-aca4-63a57deec717"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.016902 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78457c3d-d7fe-413c-aca4-63a57deec717" (UID: "78457c3d-d7fe-413c-aca4-63a57deec717"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.019850 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78457c3d-d7fe-413c-aca4-63a57deec717" (UID: "78457c3d-d7fe-413c-aca4-63a57deec717"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.040336 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78457c3d-d7fe-413c-aca4-63a57deec717" (UID: "78457c3d-d7fe-413c-aca4-63a57deec717"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.046926 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-config" (OuterVolumeSpecName: "config") pod "78457c3d-d7fe-413c-aca4-63a57deec717" (UID: "78457c3d-d7fe-413c-aca4-63a57deec717"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056678 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzxwl\" (UniqueName: \"kubernetes.io/projected/78457c3d-d7fe-413c-aca4-63a57deec717-kube-api-access-lzxwl\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056707 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056721 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056731 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056744 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d63d49d-2614-4329-abde-118b126b9a0d-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056757 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056767 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.056777 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78457c3d-d7fe-413c-aca4-63a57deec717-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.099074 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.195798 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dbdc9c48f-dtqv6"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.202678 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9qtfk" event={"ID":"6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b","Type":"ContainerDied","Data":"c7a5f9ed3a25cb79040d034f04a86b18dd17953bc1f25331620cb4bc51b85287"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.202723 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a5f9ed3a25cb79040d034f04a86b18dd17953bc1f25331620cb4bc51b85287" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.202735 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9qtfk" Dec 02 10:39:05 crc kubenswrapper[4679]: W1202 10:39:05.202854 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeecb1bea_c78b_4888_b9a8_e868ab82a9db.slice/crio-aa4f576b5db4382d8333c3f1f42c6dad094bad4f58311a6dabb01e8cbd8074f0 WatchSource:0}: Error finding container aa4f576b5db4382d8333c3f1f42c6dad094bad4f58311a6dabb01e8cbd8074f0: Status 404 returned error can't find the container with id aa4f576b5db4382d8333c3f1f42c6dad094bad4f58311a6dabb01e8cbd8074f0 Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.204673 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3d63d49d-2614-4329-abde-118b126b9a0d","Type":"ContainerDied","Data":"ab0f364315ee7a6be3b1eb37dcdc12a92412b89ac227b3f095cb544cb5602e1d"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.204708 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.204714 4679 scope.go:117] "RemoveContainer" containerID="16a16785b1a0ad37fb80eaba358e35f064a1f75277a951eab1e2d28a63824d75" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.211319 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5799263-7d7f-4195-b6b2-70d9315e3a3f","Type":"ContainerStarted","Data":"869d8d431496fb084a9e88b79149dd26f397905071c2589db2c265362644bb68"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.213892 4679 generic.go:334] "Generic (PLEG): container finished" podID="78457c3d-d7fe-413c-aca4-63a57deec717" containerID="5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9" exitCode=0 Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.213948 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" event={"ID":"78457c3d-d7fe-413c-aca4-63a57deec717","Type":"ContainerDied","Data":"5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.213973 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" event={"ID":"78457c3d-d7fe-413c-aca4-63a57deec717","Type":"ContainerDied","Data":"27d6a3d032f67c7079cecbd807805066d8af5043e67eea980826575cb7272ce2"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.214042 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jtqff" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.215494 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"470284bf-e122-400f-9722-f8d4f23a75cf","Type":"ContainerStarted","Data":"9d250a50918656caa7bccc420775be160e14bf433e82b6f0fe2b80dde3e48a29"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.218982 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-9kjcb" event={"ID":"055921c7-b108-40a9-a4da-2947ca27c2d8","Type":"ContainerDied","Data":"c0305d18386fa0d9f6702fbe82910fca2bec0b322a2a701a05648bf27a72a4a2"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.219022 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0305d18386fa0d9f6702fbe82910fca2bec0b322a2a701a05648bf27a72a4a2" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.219087 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-9kjcb" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.221910 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-86lxg" event={"ID":"f071f49e-c3e2-4919-a930-dd5fd654cf55","Type":"ContainerDied","Data":"602b714ce0dc4f36eaf21508a9a0d1b10fc2be668769576e7d64ab13a3b1d1d8"} Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.221938 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="602b714ce0dc4f36eaf21508a9a0d1b10fc2be668769576e7d64ab13a3b1d1d8" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.221962 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-86lxg" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.237732 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.247740 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.259249 4679 scope.go:117] "RemoveContainer" containerID="fbb27f2ae7c06d8c39442eafd4cc9b95d377a64f8a448fb9e9ad49970b9d4a05" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274327 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274777 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055921c7-b108-40a9-a4da-2947ca27c2d8" containerName="placement-db-sync" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274789 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="055921c7-b108-40a9-a4da-2947ca27c2d8" containerName="placement-db-sync" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274800 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-httpd" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274807 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-httpd" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274821 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-log" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274829 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-log" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274848 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" containerName="dnsmasq-dns" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274856 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" containerName="dnsmasq-dns" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274863 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f071f49e-c3e2-4919-a930-dd5fd654cf55" containerName="barbican-db-sync" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274869 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f071f49e-c3e2-4919-a930-dd5fd654cf55" containerName="barbican-db-sync" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274876 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" containerName="keystone-bootstrap" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274883 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" containerName="keystone-bootstrap" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.274894 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" containerName="init" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.274899 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" containerName="init" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.275175 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-log" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.275191 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" containerName="glance-httpd" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.275198 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" containerName="keystone-bootstrap" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.275206 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" containerName="dnsmasq-dns" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.275222 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="055921c7-b108-40a9-a4da-2947ca27c2d8" containerName="placement-db-sync" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.275232 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f071f49e-c3e2-4919-a930-dd5fd654cf55" containerName="barbican-db-sync" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.276366 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.278793 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.279026 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.280884 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jtqff"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.297025 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.297466 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.314385 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jtqff"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.324190 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.339547 4679 scope.go:117] "RemoveContainer" containerID="5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.379645 4679 scope.go:117] "RemoveContainer" containerID="3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.393401 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.393451 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.462517 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.462568 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.462602 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.462726 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zz4p\" (UniqueName: \"kubernetes.io/projected/8fae901a-a069-4931-adad-b71b0cd4df8e-kube-api-access-6zz4p\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.462861 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-logs\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.463107 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.463255 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.463476 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.491597 4679 scope.go:117] "RemoveContainer" containerID="5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.500783 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9\": container with ID starting with 5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9 not found: ID does not exist" containerID="5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.500829 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9"} err="failed to get container status \"5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9\": rpc error: code = NotFound desc = could not find container \"5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9\": container with ID starting with 5eddeed52700e19d4649011470a5ebce2e8ab210282d7ab1dda2a5e6c2ad07e9 not found: ID does not exist" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.500861 4679 scope.go:117] "RemoveContainer" containerID="3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780" Dec 02 10:39:05 crc kubenswrapper[4679]: E1202 10:39:05.501204 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780\": container with ID starting with 3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780 not found: ID does not exist" containerID="3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.501229 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780"} err="failed to get container status \"3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780\": rpc error: code = NotFound desc = could not find container \"3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780\": container with ID starting with 3034f54bc92fcef98fe062db10550a9c9272970db1f0760787870a7a66b56780 not found: ID does not exist" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.564999 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565082 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565154 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565215 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565233 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565251 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565286 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zz4p\" (UniqueName: \"kubernetes.io/projected/8fae901a-a069-4931-adad-b71b0cd4df8e-kube-api-access-6zz4p\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.565500 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-logs\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.566484 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.566694 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.571446 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.572043 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-logs\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.578538 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.589077 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.590244 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.595025 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zz4p\" (UniqueName: \"kubernetes.io/projected/8fae901a-a069-4931-adad-b71b0cd4df8e-kube-api-access-6zz4p\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.611649 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.771604 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-947c4b5c6-7fq4t"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.773582 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.778555 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.783326 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.783408 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.783628 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-jqtrr" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.783748 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.806293 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-947c4b5c6-7fq4t"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.871522 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f15bf3-d397-4937-bcd3-c61999ce8dae-logs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.871936 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6kgt\" (UniqueName: \"kubernetes.io/projected/19f15bf3-d397-4937-bcd3-c61999ce8dae-kube-api-access-p6kgt\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.871987 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-public-tls-certs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.872113 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-scripts\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.872169 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-config-data\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.873409 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-combined-ca-bundle\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.873472 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-internal-tls-certs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.892359 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6558c68964-dhdww"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.893587 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.905888 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.906060 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.906165 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.906271 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.906417 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5x6pg" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.906827 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.912749 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.985290 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f15bf3-d397-4937-bcd3-c61999ce8dae-logs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.985541 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6kgt\" (UniqueName: \"kubernetes.io/projected/19f15bf3-d397-4937-bcd3-c61999ce8dae-kube-api-access-p6kgt\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.985643 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-public-tls-certs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.985838 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-scripts\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.985942 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-config-data\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.986065 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-combined-ca-bundle\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.986143 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-internal-tls-certs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.987064 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19f15bf3-d397-4937-bcd3-c61999ce8dae-logs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.992372 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6558c68964-dhdww"] Dec 02 10:39:05 crc kubenswrapper[4679]: I1202 10:39:05.996032 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-public-tls-certs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.006740 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-scripts\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.007272 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-config-data\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.017538 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-combined-ca-bundle\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.035944 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6kgt\" (UniqueName: \"kubernetes.io/projected/19f15bf3-d397-4937-bcd3-c61999ce8dae-kube-api-access-p6kgt\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.037601 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f15bf3-d397-4937-bcd3-c61999ce8dae-internal-tls-certs\") pod \"placement-947c4b5c6-7fq4t\" (UID: \"19f15bf3-d397-4937-bcd3-c61999ce8dae\") " pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.062283 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-9dcfb6847-f2pxh"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.066984 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.081250 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-b8zhd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.081547 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.081661 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.091918 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-public-tls-certs\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092025 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-credential-keys\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092335 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-internal-tls-certs\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092382 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-scripts\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092519 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-config-data\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092761 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-combined-ca-bundle\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092820 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwkmw\" (UniqueName: \"kubernetes.io/projected/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-kube-api-access-nwkmw\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.092863 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-fernet-keys\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.099837 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.109408 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9dcfb6847-f2pxh"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.222421 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-combined-ca-bundle\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.222915 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-internal-tls-certs\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.222957 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-scripts\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.223133 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-config-data\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.223454 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b25066f-17f7-42c3-8448-a25ff170388a-logs\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.224689 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-combined-ca-bundle\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.224773 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwkmw\" (UniqueName: \"kubernetes.io/projected/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-kube-api-access-nwkmw\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.224810 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-fernet-keys\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.224859 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb8gm\" (UniqueName: \"kubernetes.io/projected/5b25066f-17f7-42c3-8448-a25ff170388a-kube-api-access-qb8gm\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.224886 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-config-data\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.224947 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-public-tls-certs\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.227047 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-credential-keys\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.227177 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-config-data-custom\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.275094 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-fernet-keys\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.284425 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-internal-tls-certs\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.293628 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-public-tls-certs\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.304268 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-config-data\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.310496 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-combined-ca-bundle\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.315091 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-scripts\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.315564 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-credential-keys\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.330776 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwkmw\" (UniqueName: \"kubernetes.io/projected/6691f3fc-9b58-4cc5-af4e-b4aaf10036e9-kube-api-access-nwkmw\") pod \"keystone-6558c68964-dhdww\" (UID: \"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9\") " pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.363557 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dbdc9c48f-dtqv6" event={"ID":"eecb1bea-c78b-4888-b9a8-e868ab82a9db","Type":"ContainerStarted","Data":"0bb3cb11b97af948d170c591ae3b9db21582684a8266cc070fe7f91ef8b54285"} Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.366901 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dbdc9c48f-dtqv6" event={"ID":"eecb1bea-c78b-4888-b9a8-e868ab82a9db","Type":"ContainerStarted","Data":"342920947dcadea6e8e94c3d16ac791a294ee5ef200540985bfa20aa3c5871a0"} Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.366920 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dbdc9c48f-dtqv6" event={"ID":"eecb1bea-c78b-4888-b9a8-e868ab82a9db","Type":"ContainerStarted","Data":"aa4f576b5db4382d8333c3f1f42c6dad094bad4f58311a6dabb01e8cbd8074f0"} Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.367804 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.383617 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-config-data-custom\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.383665 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-combined-ca-bundle\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.383769 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b25066f-17f7-42c3-8448-a25ff170388a-logs\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.383806 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb8gm\" (UniqueName: \"kubernetes.io/projected/5b25066f-17f7-42c3-8448-a25ff170388a-kube-api-access-qb8gm\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.383823 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-config-data\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.404474 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b25066f-17f7-42c3-8448-a25ff170388a-logs\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.408127 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vggwm"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.413523 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.440841 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-config-data-custom\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.444469 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-combined-ca-bundle\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.446011 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b25066f-17f7-42c3-8448-a25ff170388a-config-data\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.459019 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5799263-7d7f-4195-b6b2-70d9315e3a3f","Type":"ContainerStarted","Data":"a2c1e359e1d0038891a7043726146cd0875dec70ec23bd403e0a37eec1b35286"} Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.465881 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb8gm\" (UniqueName: \"kubernetes.io/projected/5b25066f-17f7-42c3-8448-a25ff170388a-kube-api-access-qb8gm\") pod \"barbican-worker-9dcfb6847-f2pxh\" (UID: \"5b25066f-17f7-42c3-8448-a25ff170388a\") " pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.476925 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-78df7f5646-r7nmv"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.478624 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.482785 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.485607 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.485708 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.485728 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-config\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.485796 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.485810 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.485856 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7c82\" (UniqueName: \"kubernetes.io/projected/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-kube-api-access-m7c82\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.502458 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vggwm"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.512611 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-78df7f5646-r7nmv"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.548146 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598740 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7c82\" (UniqueName: \"kubernetes.io/projected/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-kube-api-access-m7c82\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598803 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6f2d\" (UniqueName: \"kubernetes.io/projected/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-kube-api-access-l6f2d\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598850 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-config-data-custom\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598881 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598900 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-logs\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598946 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-combined-ca-bundle\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598968 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-config-data\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.598986 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.599004 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-config\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.599040 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.599057 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.599976 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.600245 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.600897 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.601020 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-config\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.601200 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.633188 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7c82\" (UniqueName: \"kubernetes.io/projected/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-kube-api-access-m7c82\") pod \"dnsmasq-dns-75c8ddd69c-vggwm\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.643658 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85c6568ff4-fjmqd"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.645719 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.649578 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.658682 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85c6568ff4-fjmqd"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.664936 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6dbdc9c48f-dtqv6" podStartSLOduration=7.664912519 podStartE2EDuration="7.664912519s" podCreationTimestamp="2025-12-02 10:38:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:06.446499058 +0000 UTC m=+1199.776637938" watchObservedRunningTime="2025-12-02 10:39:06.664912519 +0000 UTC m=+1199.995051379" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706652 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706730 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ggv6\" (UniqueName: \"kubernetes.io/projected/ed234cbf-40c8-4856-bad3-6bf38c7ff610-kube-api-access-9ggv6\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706759 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6f2d\" (UniqueName: \"kubernetes.io/projected/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-kube-api-access-l6f2d\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706851 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data-custom\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706892 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-config-data-custom\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706927 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-logs\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.706999 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-combined-ca-bundle\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.707022 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-config-data\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.707083 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-combined-ca-bundle\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.707113 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed234cbf-40c8-4856-bad3-6bf38c7ff610-logs\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.709251 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-logs\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.714186 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-config-data-custom\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.723325 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-combined-ca-bundle\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.723721 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-config-data\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.738266 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6f2d\" (UniqueName: \"kubernetes.io/projected/3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f-kube-api-access-l6f2d\") pod \"barbican-keystone-listener-78df7f5646-r7nmv\" (UID: \"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f\") " pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.743758 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.761032 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9dcfb6847-f2pxh" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.809294 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-combined-ca-bundle\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.809376 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed234cbf-40c8-4856-bad3-6bf38c7ff610-logs\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.809413 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.809449 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ggv6\" (UniqueName: \"kubernetes.io/projected/ed234cbf-40c8-4856-bad3-6bf38c7ff610-kube-api-access-9ggv6\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.809503 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data-custom\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.810436 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed234cbf-40c8-4856-bad3-6bf38c7ff610-logs\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.816874 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data-custom\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.817729 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.824984 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-combined-ca-bundle\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.835038 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.836879 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ggv6\" (UniqueName: \"kubernetes.io/projected/ed234cbf-40c8-4856-bad3-6bf38c7ff610-kube-api-access-9ggv6\") pod \"barbican-api-85c6568ff4-fjmqd\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.855551 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" Dec 02 10:39:06 crc kubenswrapper[4679]: I1202 10:39:06.984369 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.028408 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d63d49d-2614-4329-abde-118b126b9a0d" path="/var/lib/kubelet/pods/3d63d49d-2614-4329-abde-118b126b9a0d/volumes" Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.029820 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78457c3d-d7fe-413c-aca4-63a57deec717" path="/var/lib/kubelet/pods/78457c3d-d7fe-413c-aca4-63a57deec717/volumes" Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.030833 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-947c4b5c6-7fq4t"] Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.051740 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6558c68964-dhdww"] Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.461747 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9dcfb6847-f2pxh"] Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.548912 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-947c4b5c6-7fq4t" event={"ID":"19f15bf3-d397-4937-bcd3-c61999ce8dae","Type":"ContainerStarted","Data":"19dec9d6a5827e00db90d4fe4cfb87568b4a3cbdefa62481286259b92ce09228"} Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.582529 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6558c68964-dhdww" event={"ID":"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9","Type":"ContainerStarted","Data":"1798fe9186de46aab4611d77c7d26973b1a4d11d4647c7b4b0d07981ab6c2412"} Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.585457 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8fae901a-a069-4931-adad-b71b0cd4df8e","Type":"ContainerStarted","Data":"91e90640aede6ecb4361843fe8885180cb2b947b377cee9304f7db57be2d6173"} Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.638098 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vggwm"] Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.735522 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-78df7f5646-r7nmv"] Dec 02 10:39:07 crc kubenswrapper[4679]: I1202 10:39:07.856469 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85c6568ff4-fjmqd"] Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.620214 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85c6568ff4-fjmqd" event={"ID":"ed234cbf-40c8-4856-bad3-6bf38c7ff610","Type":"ContainerStarted","Data":"e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.620856 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85c6568ff4-fjmqd" event={"ID":"ed234cbf-40c8-4856-bad3-6bf38c7ff610","Type":"ContainerStarted","Data":"a890f66ad3f306f4d21f9097172fe9cb20dc9cf6d1c1064aad408db869c04fe9"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.622408 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9dcfb6847-f2pxh" event={"ID":"5b25066f-17f7-42c3-8448-a25ff170388a","Type":"ContainerStarted","Data":"59f49cb9dbc3fbd1e028be49d5bb879701b28ae842a17c0a687976e08b78be6a"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.624146 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-947c4b5c6-7fq4t" event={"ID":"19f15bf3-d397-4937-bcd3-c61999ce8dae","Type":"ContainerStarted","Data":"e3c660c29402bb3f547765f5456905a0bb795a6589827b017d22fbe8217294df"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.625424 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6558c68964-dhdww" event={"ID":"6691f3fc-9b58-4cc5-af4e-b4aaf10036e9","Type":"ContainerStarted","Data":"5e9813e19103f316c32cd34c88827e01d54586f890e253871ea32080e7a5f35b"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.626845 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.627892 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" event={"ID":"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f","Type":"ContainerStarted","Data":"d6841af00d88de8806e40929576219d1a2beba0050034f362e5e71771f663cb6"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.631959 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5799263-7d7f-4195-b6b2-70d9315e3a3f","Type":"ContainerStarted","Data":"8c7f669e05352f9712222cdbe0ad49be5385a980e06c61e53e2a95a934486904"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.651529 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6558c68964-dhdww" podStartSLOduration=3.651508198 podStartE2EDuration="3.651508198s" podCreationTimestamp="2025-12-02 10:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:08.639735689 +0000 UTC m=+1201.969874549" watchObservedRunningTime="2025-12-02 10:39:08.651508198 +0000 UTC m=+1201.981647048" Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.652961 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8fae901a-a069-4931-adad-b71b0cd4df8e","Type":"ContainerStarted","Data":"b87d66c33269519376e092d1b0bf22a0441cc598188d0a792f9a159942fc33bc"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.659219 4679 generic.go:334] "Generic (PLEG): container finished" podID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerID="16c9ce7605d2726c95ebb815b0034375cb0e3f2330bfb6f4df9096a1f428fa2f" exitCode=0 Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.659261 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" event={"ID":"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce","Type":"ContainerDied","Data":"16c9ce7605d2726c95ebb815b0034375cb0e3f2330bfb6f4df9096a1f428fa2f"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.659283 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" event={"ID":"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce","Type":"ContainerStarted","Data":"863eece2e504fb8187ab2245295d21507dc65a46b1c4672133d7dbd7ed39c958"} Dec 02 10:39:08 crc kubenswrapper[4679]: I1202 10:39:08.676299 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.676278783 podStartE2EDuration="6.676278783s" podCreationTimestamp="2025-12-02 10:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:08.672845124 +0000 UTC m=+1202.002983984" watchObservedRunningTime="2025-12-02 10:39:08.676278783 +0000 UTC m=+1202.006417633" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.046052 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5fd9984b94-spfgc"] Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.047771 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.051596 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.052179 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.077222 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fd9984b94-spfgc"] Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.099794 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-config-data\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.100209 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-internal-tls-certs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.100236 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/028dd5f7-9b68-45af-8515-abef1cc975d0-logs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.100272 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnqrv\" (UniqueName: \"kubernetes.io/projected/028dd5f7-9b68-45af-8515-abef1cc975d0-kube-api-access-dnqrv\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.100328 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-public-tls-certs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.100414 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-combined-ca-bundle\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.100477 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-config-data-custom\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201669 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-combined-ca-bundle\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201748 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-config-data-custom\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201786 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-config-data\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201810 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-internal-tls-certs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201825 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/028dd5f7-9b68-45af-8515-abef1cc975d0-logs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201858 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnqrv\" (UniqueName: \"kubernetes.io/projected/028dd5f7-9b68-45af-8515-abef1cc975d0-kube-api-access-dnqrv\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.201887 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-public-tls-certs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.214013 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-combined-ca-bundle\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.215746 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/028dd5f7-9b68-45af-8515-abef1cc975d0-logs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.217104 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-public-tls-certs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.218031 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-config-data-custom\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.219923 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-internal-tls-certs\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.223789 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/028dd5f7-9b68-45af-8515-abef1cc975d0-config-data\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.235876 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnqrv\" (UniqueName: \"kubernetes.io/projected/028dd5f7-9b68-45af-8515-abef1cc975d0-kube-api-access-dnqrv\") pod \"barbican-api-5fd9984b94-spfgc\" (UID: \"028dd5f7-9b68-45af-8515-abef1cc975d0\") " pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.398593 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.691996 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" event={"ID":"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce","Type":"ContainerStarted","Data":"40bd0ffa15a8f1c691c47a4c9d9e744e3a368de748fe3021427a268baebdb592"} Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.692394 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.698042 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85c6568ff4-fjmqd" event={"ID":"ed234cbf-40c8-4856-bad3-6bf38c7ff610","Type":"ContainerStarted","Data":"1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565"} Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.699016 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.699059 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.716605 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" podStartSLOduration=4.716583774 podStartE2EDuration="4.716583774s" podCreationTimestamp="2025-12-02 10:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:09.710928001 +0000 UTC m=+1203.041066871" watchObservedRunningTime="2025-12-02 10:39:09.716583774 +0000 UTC m=+1203.046722634" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.718465 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-947c4b5c6-7fq4t" event={"ID":"19f15bf3-d397-4937-bcd3-c61999ce8dae","Type":"ContainerStarted","Data":"7bb05f2b82ebce63a5439c3e06716db0b6b4b3bc2b54b7f595ff64905a4b75ff"} Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.719346 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.739648 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85c6568ff4-fjmqd" podStartSLOduration=3.7396335990000003 podStartE2EDuration="3.739633599s" podCreationTimestamp="2025-12-02 10:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:09.737456466 +0000 UTC m=+1203.067595326" watchObservedRunningTime="2025-12-02 10:39:09.739633599 +0000 UTC m=+1203.069772449" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.776044 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-947c4b5c6-7fq4t" podStartSLOduration=4.776022569 podStartE2EDuration="4.776022569s" podCreationTimestamp="2025-12-02 10:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:09.775118263 +0000 UTC m=+1203.105257123" watchObservedRunningTime="2025-12-02 10:39:09.776022569 +0000 UTC m=+1203.106161419" Dec 02 10:39:09 crc kubenswrapper[4679]: I1202 10:39:09.929402 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fd9984b94-spfgc"] Dec 02 10:39:10 crc kubenswrapper[4679]: W1202 10:39:10.323035 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod028dd5f7_9b68_45af_8515_abef1cc975d0.slice/crio-282379bf0f424bd20852be2ba30b40b95e77f46a51d2da65d699aaa38f4dfc4e WatchSource:0}: Error finding container 282379bf0f424bd20852be2ba30b40b95e77f46a51d2da65d699aaa38f4dfc4e: Status 404 returned error can't find the container with id 282379bf0f424bd20852be2ba30b40b95e77f46a51d2da65d699aaa38f4dfc4e Dec 02 10:39:10 crc kubenswrapper[4679]: I1202 10:39:10.727797 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fd9984b94-spfgc" event={"ID":"028dd5f7-9b68-45af-8515-abef1cc975d0","Type":"ContainerStarted","Data":"282379bf0f424bd20852be2ba30b40b95e77f46a51d2da65d699aaa38f4dfc4e"} Dec 02 10:39:10 crc kubenswrapper[4679]: I1202 10:39:10.729787 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8fae901a-a069-4931-adad-b71b0cd4df8e","Type":"ContainerStarted","Data":"b36b75676ab9c393bc54e903d245097061243d086034ab24a6f56ca3f1bebd72"} Dec 02 10:39:10 crc kubenswrapper[4679]: I1202 10:39:10.734856 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqb7p" event={"ID":"048db06f-54db-4862-b8c8-b7e9cbb55095","Type":"ContainerStarted","Data":"63c81f5df5186090ed9c4c7eeaa433b87cd234a0fd8ab2ba4c543da186a82336"} Dec 02 10:39:10 crc kubenswrapper[4679]: I1202 10:39:10.735964 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:10 crc kubenswrapper[4679]: I1202 10:39:10.754067 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.754047783 podStartE2EDuration="5.754047783s" podCreationTimestamp="2025-12-02 10:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:10.745865187 +0000 UTC m=+1204.076004047" watchObservedRunningTime="2025-12-02 10:39:10.754047783 +0000 UTC m=+1204.084186643" Dec 02 10:39:10 crc kubenswrapper[4679]: I1202 10:39:10.774261 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-bqb7p" podStartSLOduration=3.99907478 podStartE2EDuration="54.774240155s" podCreationTimestamp="2025-12-02 10:38:16 +0000 UTC" firstStartedPulling="2025-12-02 10:38:17.785301437 +0000 UTC m=+1151.115440297" lastFinishedPulling="2025-12-02 10:39:08.560466812 +0000 UTC m=+1201.890605672" observedRunningTime="2025-12-02 10:39:10.771076594 +0000 UTC m=+1204.101215464" watchObservedRunningTime="2025-12-02 10:39:10.774240155 +0000 UTC m=+1204.104379035" Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.747331 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fd9984b94-spfgc" event={"ID":"028dd5f7-9b68-45af-8515-abef1cc975d0","Type":"ContainerStarted","Data":"2824046c627b375df962cd62e04efa57f5b797d7601a66915d5783c6c731b9d1"} Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.747391 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fd9984b94-spfgc" event={"ID":"028dd5f7-9b68-45af-8515-abef1cc975d0","Type":"ContainerStarted","Data":"3090b3241cac6243609307d650a14990258f7e8f49efb32ddc70a2f29032ba05"} Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.748573 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.748613 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.763668 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" event={"ID":"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f","Type":"ContainerStarted","Data":"70b67cbbf46a6730ba12efe928c0957d734f05718d80999f5375e7dc3de3f7d0"} Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.763723 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" event={"ID":"3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f","Type":"ContainerStarted","Data":"0f3d3bf33f4ded7b8963b618e65f9bcdff56797e8350c4016b6693dd10efe016"} Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.779377 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5fd9984b94-spfgc" podStartSLOduration=2.779352991 podStartE2EDuration="2.779352991s" podCreationTimestamp="2025-12-02 10:39:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:11.76440421 +0000 UTC m=+1205.094543060" watchObservedRunningTime="2025-12-02 10:39:11.779352991 +0000 UTC m=+1205.109491851" Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.788794 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9dcfb6847-f2pxh" event={"ID":"5b25066f-17f7-42c3-8448-a25ff170388a","Type":"ContainerStarted","Data":"93d867173625f563bedd75b143c2f664d0d405466b7573a0b70f682c35422790"} Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.788830 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9dcfb6847-f2pxh" event={"ID":"5b25066f-17f7-42c3-8448-a25ff170388a","Type":"ContainerStarted","Data":"0c35698600d1ed52eff95befbec7825613a272c765c76821e0db2a756975c7f3"} Dec 02 10:39:11 crc kubenswrapper[4679]: I1202 10:39:11.810576 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-78df7f5646-r7nmv" podStartSLOduration=2.7314164229999998 podStartE2EDuration="5.810557481s" podCreationTimestamp="2025-12-02 10:39:06 +0000 UTC" firstStartedPulling="2025-12-02 10:39:07.858959544 +0000 UTC m=+1201.189098424" lastFinishedPulling="2025-12-02 10:39:10.938100622 +0000 UTC m=+1204.268239482" observedRunningTime="2025-12-02 10:39:11.799132452 +0000 UTC m=+1205.129271312" watchObservedRunningTime="2025-12-02 10:39:11.810557481 +0000 UTC m=+1205.140696341" Dec 02 10:39:12 crc kubenswrapper[4679]: I1202 10:39:12.922064 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 10:39:12 crc kubenswrapper[4679]: I1202 10:39:12.922512 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 10:39:12 crc kubenswrapper[4679]: I1202 10:39:12.957116 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 10:39:12 crc kubenswrapper[4679]: I1202 10:39:12.969504 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 10:39:12 crc kubenswrapper[4679]: I1202 10:39:12.995252 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-9dcfb6847-f2pxh" podStartSLOduration=4.5715258290000005 podStartE2EDuration="7.995227928s" podCreationTimestamp="2025-12-02 10:39:05 +0000 UTC" firstStartedPulling="2025-12-02 10:39:07.515516146 +0000 UTC m=+1200.845655006" lastFinishedPulling="2025-12-02 10:39:10.939218245 +0000 UTC m=+1204.269357105" observedRunningTime="2025-12-02 10:39:11.837632003 +0000 UTC m=+1205.167770863" watchObservedRunningTime="2025-12-02 10:39:12.995227928 +0000 UTC m=+1206.325366788" Dec 02 10:39:13 crc kubenswrapper[4679]: I1202 10:39:13.811429 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 10:39:13 crc kubenswrapper[4679]: I1202 10:39:13.811472 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 10:39:13 crc kubenswrapper[4679]: I1202 10:39:13.966882 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.062619 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-947c4b5c6-7fq4t" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.301738 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bd7d57448-84dtd" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.394807 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6fd99d7d5b-sb96h" podUID="4dd62da9-d481-4fc3-bbe6-108a54d98154" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.826947 4679 generic.go:334] "Generic (PLEG): container finished" podID="048db06f-54db-4862-b8c8-b7e9cbb55095" containerID="63c81f5df5186090ed9c4c7eeaa433b87cd234a0fd8ab2ba4c543da186a82336" exitCode=0 Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.827372 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.827382 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.828090 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqb7p" event={"ID":"048db06f-54db-4862-b8c8-b7e9cbb55095","Type":"ContainerDied","Data":"63c81f5df5186090ed9c4c7eeaa433b87cd234a0fd8ab2ba4c543da186a82336"} Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.830857 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.914492 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.914540 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.949721 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:15 crc kubenswrapper[4679]: I1202 10:39:15.965808 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.292058 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.836441 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.838265 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.838295 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.932429 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.932913 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.955723 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-cltpj"] Dec 02 10:39:16 crc kubenswrapper[4679]: I1202 10:39:16.955972 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="dnsmasq-dns" containerID="cri-o://00ff65400adb4603bece89da8eaf7863189ced756ba99423db2081b70e9255c4" gracePeriod=10 Dec 02 10:39:17 crc kubenswrapper[4679]: I1202 10:39:17.882038 4679 generic.go:334] "Generic (PLEG): container finished" podID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerID="00ff65400adb4603bece89da8eaf7863189ced756ba99423db2081b70e9255c4" exitCode=0 Dec 02 10:39:17 crc kubenswrapper[4679]: I1202 10:39:17.882387 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" event={"ID":"f1884c8a-197d-4631-a7bb-4fceb1f27649","Type":"ContainerDied","Data":"00ff65400adb4603bece89da8eaf7863189ced756ba99423db2081b70e9255c4"} Dec 02 10:39:18 crc kubenswrapper[4679]: I1202 10:39:18.891849 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:39:18 crc kubenswrapper[4679]: I1202 10:39:18.892459 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.574857 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.580646 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.630899 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.660171 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-swift-storage-0\") pod \"f1884c8a-197d-4631-a7bb-4fceb1f27649\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.660953 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-config-data\") pod \"048db06f-54db-4862-b8c8-b7e9cbb55095\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.660991 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-db-sync-config-data\") pod \"048db06f-54db-4862-b8c8-b7e9cbb55095\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661014 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ptzq\" (UniqueName: \"kubernetes.io/projected/048db06f-54db-4862-b8c8-b7e9cbb55095-kube-api-access-8ptzq\") pod \"048db06f-54db-4862-b8c8-b7e9cbb55095\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661036 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-scripts\") pod \"048db06f-54db-4862-b8c8-b7e9cbb55095\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661096 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sfvv\" (UniqueName: \"kubernetes.io/projected/f1884c8a-197d-4631-a7bb-4fceb1f27649-kube-api-access-2sfvv\") pod \"f1884c8a-197d-4631-a7bb-4fceb1f27649\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661147 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-nb\") pod \"f1884c8a-197d-4631-a7bb-4fceb1f27649\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661185 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-sb\") pod \"f1884c8a-197d-4631-a7bb-4fceb1f27649\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661206 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-svc\") pod \"f1884c8a-197d-4631-a7bb-4fceb1f27649\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661246 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/048db06f-54db-4862-b8c8-b7e9cbb55095-etc-machine-id\") pod \"048db06f-54db-4862-b8c8-b7e9cbb55095\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661331 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-config\") pod \"f1884c8a-197d-4631-a7bb-4fceb1f27649\" (UID: \"f1884c8a-197d-4631-a7bb-4fceb1f27649\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.661363 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-combined-ca-bundle\") pod \"048db06f-54db-4862-b8c8-b7e9cbb55095\" (UID: \"048db06f-54db-4862-b8c8-b7e9cbb55095\") " Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.674420 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/048db06f-54db-4862-b8c8-b7e9cbb55095-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "048db06f-54db-4862-b8c8-b7e9cbb55095" (UID: "048db06f-54db-4862-b8c8-b7e9cbb55095"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.688494 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "048db06f-54db-4862-b8c8-b7e9cbb55095" (UID: "048db06f-54db-4862-b8c8-b7e9cbb55095"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.690549 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048db06f-54db-4862-b8c8-b7e9cbb55095-kube-api-access-8ptzq" (OuterVolumeSpecName: "kube-api-access-8ptzq") pod "048db06f-54db-4862-b8c8-b7e9cbb55095" (UID: "048db06f-54db-4862-b8c8-b7e9cbb55095"). InnerVolumeSpecName "kube-api-access-8ptzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.709648 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1884c8a-197d-4631-a7bb-4fceb1f27649-kube-api-access-2sfvv" (OuterVolumeSpecName: "kube-api-access-2sfvv") pod "f1884c8a-197d-4631-a7bb-4fceb1f27649" (UID: "f1884c8a-197d-4631-a7bb-4fceb1f27649"). InnerVolumeSpecName "kube-api-access-2sfvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.713420 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-scripts" (OuterVolumeSpecName: "scripts") pod "048db06f-54db-4862-b8c8-b7e9cbb55095" (UID: "048db06f-54db-4862-b8c8-b7e9cbb55095"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.747845 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "048db06f-54db-4862-b8c8-b7e9cbb55095" (UID: "048db06f-54db-4862-b8c8-b7e9cbb55095"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.767633 4679 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.767855 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ptzq\" (UniqueName: \"kubernetes.io/projected/048db06f-54db-4862-b8c8-b7e9cbb55095-kube-api-access-8ptzq\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.767868 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.767877 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sfvv\" (UniqueName: \"kubernetes.io/projected/f1884c8a-197d-4631-a7bb-4fceb1f27649-kube-api-access-2sfvv\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.767886 4679 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/048db06f-54db-4862-b8c8-b7e9cbb55095-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.767894 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.780913 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f1884c8a-197d-4631-a7bb-4fceb1f27649" (UID: "f1884c8a-197d-4631-a7bb-4fceb1f27649"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.784398 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-config-data" (OuterVolumeSpecName: "config-data") pod "048db06f-54db-4862-b8c8-b7e9cbb55095" (UID: "048db06f-54db-4862-b8c8-b7e9cbb55095"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.794251 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1884c8a-197d-4631-a7bb-4fceb1f27649" (UID: "f1884c8a-197d-4631-a7bb-4fceb1f27649"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.799750 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f1884c8a-197d-4631-a7bb-4fceb1f27649" (UID: "f1884c8a-197d-4631-a7bb-4fceb1f27649"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.838209 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f1884c8a-197d-4631-a7bb-4fceb1f27649" (UID: "f1884c8a-197d-4631-a7bb-4fceb1f27649"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.851937 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-config" (OuterVolumeSpecName: "config") pod "f1884c8a-197d-4631-a7bb-4fceb1f27649" (UID: "f1884c8a-197d-4631-a7bb-4fceb1f27649"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.871441 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.871479 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.871488 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.871498 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.871507 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1884c8a-197d-4631-a7bb-4fceb1f27649-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.871518 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/048db06f-54db-4862-b8c8-b7e9cbb55095-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.923716 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" event={"ID":"f1884c8a-197d-4631-a7bb-4fceb1f27649","Type":"ContainerDied","Data":"7a1beea65d596f38ba363116a50fce95752e03ca4845ffab5962eed607c13232"} Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.923779 4679 scope.go:117] "RemoveContainer" containerID="00ff65400adb4603bece89da8eaf7863189ced756ba99423db2081b70e9255c4" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.923929 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.946375 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-bqb7p" event={"ID":"048db06f-54db-4862-b8c8-b7e9cbb55095","Type":"ContainerDied","Data":"85fda557c9826c0ec018e32ca57299f707fbc6fa95cac2e07570b25f88150d2c"} Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.946407 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85fda557c9826c0ec018e32ca57299f707fbc6fa95cac2e07570b25f88150d2c" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.946469 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-bqb7p" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.969330 4679 scope.go:117] "RemoveContainer" containerID="335c4337cafedbf9d06c3114c91da5fa28cff05b64e159adc246b627b4961969" Dec 02 10:39:19 crc kubenswrapper[4679]: I1202 10:39:19.977064 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-cltpj"] Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.002452 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-cltpj"] Dec 02 10:39:20 crc kubenswrapper[4679]: E1202 10:39:20.026645 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"ceilometer-notification-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.318906 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.318986 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.325811 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.625010 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.871628 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:20 crc kubenswrapper[4679]: E1202 10:39:20.872172 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="dnsmasq-dns" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.872213 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="dnsmasq-dns" Dec 02 10:39:20 crc kubenswrapper[4679]: E1202 10:39:20.872233 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="init" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.872242 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="init" Dec 02 10:39:20 crc kubenswrapper[4679]: E1202 10:39:20.872259 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048db06f-54db-4862-b8c8-b7e9cbb55095" containerName="cinder-db-sync" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.872267 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="048db06f-54db-4862-b8c8-b7e9cbb55095" containerName="cinder-db-sync" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.872565 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="048db06f-54db-4862-b8c8-b7e9cbb55095" containerName="cinder-db-sync" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.872590 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="dnsmasq-dns" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.873819 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.878778 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pt7cd" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.878836 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.878802 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.879093 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 02 10:39:20 crc kubenswrapper[4679]: I1202 10:39:20.890484 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.016763 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a91754-89a4-44e6-8f0b-e511785302ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.021655 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.021771 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.021814 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.021940 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.022105 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zbfl\" (UniqueName: \"kubernetes.io/projected/79a91754-89a4-44e6-8f0b-e511785302ad-kube-api-access-9zbfl\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.020737 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" path="/var/lib/kubelet/pods/f1884c8a-197d-4631-a7bb-4fceb1f27649/volumes" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.032989 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2wk9g"] Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.071718 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.091124 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2wk9g"] Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.099459 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="sg-core" containerID="cri-o://9d250a50918656caa7bccc420775be160e14bf433e82b6f0fe2b80dde3e48a29" gracePeriod=30 Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.099816 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"470284bf-e122-400f-9722-f8d4f23a75cf","Type":"ContainerStarted","Data":"28942aeb26aa591afd2839a9a64f669d09d4f73193f12bddc597c647967b3efe"} Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.099869 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.099928 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="proxy-httpd" containerID="cri-o://28942aeb26aa591afd2839a9a64f669d09d4f73193f12bddc597c647967b3efe" gracePeriod=30 Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.119420 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.121605 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.135858 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.136778 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.136855 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zbfl\" (UniqueName: \"kubernetes.io/projected/79a91754-89a4-44e6-8f0b-e511785302ad-kube-api-access-9zbfl\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.136886 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a91754-89a4-44e6-8f0b-e511785302ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.136923 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.136957 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.136974 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.137615 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a91754-89a4-44e6-8f0b-e511785302ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.155782 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.155991 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.159916 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.192711 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.193096 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.196699 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zbfl\" (UniqueName: \"kubernetes.io/projected/79a91754-89a4-44e6-8f0b-e511785302ad-kube-api-access-9zbfl\") pod \"cinder-scheduler-0\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253479 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-scripts\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253547 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253575 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnkzx\" (UniqueName: \"kubernetes.io/projected/5985d57a-3142-4a00-840e-060b8bba4e70-kube-api-access-jnkzx\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253633 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5985d57a-3142-4a00-840e-060b8bba4e70-logs\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253657 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253676 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253695 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data-custom\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253714 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds98j\" (UniqueName: \"kubernetes.io/projected/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-kube-api-access-ds98j\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253729 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5985d57a-3142-4a00-840e-060b8bba4e70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253743 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253777 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253802 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.253820 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-config\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.257231 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358452 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnkzx\" (UniqueName: \"kubernetes.io/projected/5985d57a-3142-4a00-840e-060b8bba4e70-kube-api-access-jnkzx\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358516 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5985d57a-3142-4a00-840e-060b8bba4e70-logs\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358541 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358564 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358584 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data-custom\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358601 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds98j\" (UniqueName: \"kubernetes.io/projected/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-kube-api-access-ds98j\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358619 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5985d57a-3142-4a00-840e-060b8bba4e70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358633 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358659 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358685 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358700 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-config\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358741 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-scripts\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.358779 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.359931 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.360443 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5985d57a-3142-4a00-840e-060b8bba4e70-logs\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.364810 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.365203 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.365344 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.365945 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.368877 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5985d57a-3142-4a00-840e-060b8bba4e70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.369472 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-config\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.374138 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data-custom\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.374679 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.376898 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-scripts\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.380124 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnkzx\" (UniqueName: \"kubernetes.io/projected/5985d57a-3142-4a00-840e-060b8bba4e70-kube-api-access-jnkzx\") pod \"cinder-api-0\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.396185 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds98j\" (UniqueName: \"kubernetes.io/projected/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-kube-api-access-ds98j\") pod \"dnsmasq-dns-5784cf869f-2wk9g\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.422234 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.593953 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.868621 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:21 crc kubenswrapper[4679]: I1202 10:39:21.994963 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.129464 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2wk9g"] Dec 02 10:39:22 crc kubenswrapper[4679]: W1202 10:39:22.132613 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fd5e0ee_bc7f_48de_9d1c_f15f8d98e27d.slice/crio-2456db17fccc338f646e889cacf3875f21c504beeda0eea4d64550a9424f8c52 WatchSource:0}: Error finding container 2456db17fccc338f646e889cacf3875f21c504beeda0eea4d64550a9424f8c52: Status 404 returned error can't find the container with id 2456db17fccc338f646e889cacf3875f21c504beeda0eea4d64550a9424f8c52 Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.141637 4679 generic.go:334] "Generic (PLEG): container finished" podID="470284bf-e122-400f-9722-f8d4f23a75cf" containerID="28942aeb26aa591afd2839a9a64f669d09d4f73193f12bddc597c647967b3efe" exitCode=0 Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.141675 4679 generic.go:334] "Generic (PLEG): container finished" podID="470284bf-e122-400f-9722-f8d4f23a75cf" containerID="9d250a50918656caa7bccc420775be160e14bf433e82b6f0fe2b80dde3e48a29" exitCode=2 Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.141681 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"470284bf-e122-400f-9722-f8d4f23a75cf","Type":"ContainerDied","Data":"28942aeb26aa591afd2839a9a64f669d09d4f73193f12bddc597c647967b3efe"} Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.141731 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"470284bf-e122-400f-9722-f8d4f23a75cf","Type":"ContainerDied","Data":"9d250a50918656caa7bccc420775be160e14bf433e82b6f0fe2b80dde3e48a29"} Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.160867 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79a91754-89a4-44e6-8f0b-e511785302ad","Type":"ContainerStarted","Data":"404330e1aa3b7e0399731a0e8d38288be0a26b32d057eca2dff38d70d4abfc48"} Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.196639 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296547 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb9w6\" (UniqueName: \"kubernetes.io/projected/470284bf-e122-400f-9722-f8d4f23a75cf-kube-api-access-hb9w6\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296593 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-run-httpd\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296615 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-config-data\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296653 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-log-httpd\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296690 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-scripts\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296765 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-combined-ca-bundle\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.296789 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-sg-core-conf-yaml\") pod \"470284bf-e122-400f-9722-f8d4f23a75cf\" (UID: \"470284bf-e122-400f-9722-f8d4f23a75cf\") " Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.299553 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.299780 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.308428 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/470284bf-e122-400f-9722-f8d4f23a75cf-kube-api-access-hb9w6" (OuterVolumeSpecName: "kube-api-access-hb9w6") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "kube-api-access-hb9w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.317248 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-scripts" (OuterVolumeSpecName: "scripts") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.342869 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.389449 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.395603 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.399231 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.399331 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.399391 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb9w6\" (UniqueName: \"kubernetes.io/projected/470284bf-e122-400f-9722-f8d4f23a75cf-kube-api-access-hb9w6\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.399448 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.399502 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/470284bf-e122-400f-9722-f8d4f23a75cf-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.399571 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.429982 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-config-data" (OuterVolumeSpecName: "config-data") pod "470284bf-e122-400f-9722-f8d4f23a75cf" (UID: "470284bf-e122-400f-9722-f8d4f23a75cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.501447 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470284bf-e122-400f-9722-f8d4f23a75cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.788815 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fd9984b94-spfgc" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.867824 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85c6568ff4-fjmqd"] Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.868207 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api-log" containerID="cri-o://e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883" gracePeriod=30 Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.868396 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" containerID="cri-o://1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565" gracePeriod=30 Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.884272 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.932481 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 02 10:39:22 crc kubenswrapper[4679]: I1202 10:39:22.932709 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.235078 4679 generic.go:334] "Generic (PLEG): container finished" podID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerID="e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883" exitCode=143 Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.235422 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85c6568ff4-fjmqd" event={"ID":"ed234cbf-40c8-4856-bad3-6bf38c7ff610","Type":"ContainerDied","Data":"e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883"} Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.248791 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"470284bf-e122-400f-9722-f8d4f23a75cf","Type":"ContainerDied","Data":"a1101a1e7d2098618b2e9c56637858ec4b1233f58163026c7b714b4f4c4577bf"} Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.248856 4679 scope.go:117] "RemoveContainer" containerID="28942aeb26aa591afd2839a9a64f669d09d4f73193f12bddc597c647967b3efe" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.248990 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.281517 4679 generic.go:334] "Generic (PLEG): container finished" podID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerID="b3c65d1c5056672272bb1f880561342fd3be89d0cb449d5f35a743fac78a3d33" exitCode=0 Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.281601 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" event={"ID":"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d","Type":"ContainerDied","Data":"b3c65d1c5056672272bb1f880561342fd3be89d0cb449d5f35a743fac78a3d33"} Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.281628 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" event={"ID":"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d","Type":"ContainerStarted","Data":"2456db17fccc338f646e889cacf3875f21c504beeda0eea4d64550a9424f8c52"} Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.297804 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5985d57a-3142-4a00-840e-060b8bba4e70","Type":"ContainerStarted","Data":"0490e016c39c405333e564d28772a47c9530bfccf259d219c6fb77aca413735a"} Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.383542 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.393663 4679 scope.go:117] "RemoveContainer" containerID="9d250a50918656caa7bccc420775be160e14bf433e82b6f0fe2b80dde3e48a29" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.454045 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.475718 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.509392 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:23 crc kubenswrapper[4679]: E1202 10:39:23.509843 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="proxy-httpd" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.509864 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="proxy-httpd" Dec 02 10:39:23 crc kubenswrapper[4679]: E1202 10:39:23.509879 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="sg-core" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.509887 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="sg-core" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.510104 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="sg-core" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.510127 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" containerName="proxy-httpd" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.523486 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.527051 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.548632 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.549038 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.661876 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.661943 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.661977 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-run-httpd\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.662005 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-scripts\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.662061 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtnlb\" (UniqueName: \"kubernetes.io/projected/27679d46-4a68-4372-be27-9a9ceb14ac61-kube-api-access-xtnlb\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.662118 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-config-data\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.662132 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-log-httpd\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764635 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764722 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-run-httpd\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764760 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-scripts\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764835 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtnlb\" (UniqueName: \"kubernetes.io/projected/27679d46-4a68-4372-be27-9a9ceb14ac61-kube-api-access-xtnlb\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764904 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-log-httpd\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764927 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-config-data\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.764965 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.765178 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-run-httpd\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.765473 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-log-httpd\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.768776 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.768895 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-scripts\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.778095 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.781463 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-config-data\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:23 crc kubenswrapper[4679]: I1202 10:39:23.859207 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtnlb\" (UniqueName: \"kubernetes.io/projected/27679d46-4a68-4372-be27-9a9ceb14ac61-kube-api-access-xtnlb\") pod \"ceilometer-0\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " pod="openstack/ceilometer-0" Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.011783 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.037656 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.132838 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84b966f6c9-cltpj" podUID="f1884c8a-197d-4631-a7bb-4fceb1f27649" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: i/o timeout" Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.338326 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" event={"ID":"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d","Type":"ContainerStarted","Data":"0c77363366fe5b511f74f99f79a81bd694936505421107887cbd6649a574381c"} Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.340151 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.351458 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5985d57a-3142-4a00-840e-060b8bba4e70","Type":"ContainerStarted","Data":"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae"} Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.372176 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" podStartSLOduration=4.372153544 podStartE2EDuration="4.372153544s" podCreationTimestamp="2025-12-02 10:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:24.366660525 +0000 UTC m=+1217.696799385" watchObservedRunningTime="2025-12-02 10:39:24.372153544 +0000 UTC m=+1217.702292404" Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.621212 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:24 crc kubenswrapper[4679]: I1202 10:39:24.924789 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="470284bf-e122-400f-9722-f8d4f23a75cf" path="/var/lib/kubelet/pods/470284bf-e122-400f-9722-f8d4f23a75cf/volumes" Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.362836 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5985d57a-3142-4a00-840e-060b8bba4e70","Type":"ContainerStarted","Data":"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c"} Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.363239 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api-log" containerID="cri-o://5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae" gracePeriod=30 Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.363345 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.363613 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api" containerID="cri-o://f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c" gracePeriod=30 Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.378686 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79a91754-89a4-44e6-8f0b-e511785302ad","Type":"ContainerStarted","Data":"472f686773c5600ae7db88388b27ec92d15c5cddfdcfc856dfeb8c59ee4936f6"} Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.378734 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79a91754-89a4-44e6-8f0b-e511785302ad","Type":"ContainerStarted","Data":"e16ba62420535f6092ea8d7a69351e248278777a86ae35326ef77b227a7b17a6"} Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.384457 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerStarted","Data":"5f813632e448a1275ef1a9b5a3d8a288ad450ed6371cadb94fc60058fa6357bb"} Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.397510 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.397490022 podStartE2EDuration="4.397490022s" podCreationTimestamp="2025-12-02 10:39:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:25.392412076 +0000 UTC m=+1218.722550946" watchObservedRunningTime="2025-12-02 10:39:25.397490022 +0000 UTC m=+1218.727628882" Dec 02 10:39:25 crc kubenswrapper[4679]: I1202 10:39:25.422195 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.06076771 podStartE2EDuration="5.422177784s" podCreationTimestamp="2025-12-02 10:39:20 +0000 UTC" firstStartedPulling="2025-12-02 10:39:21.884103557 +0000 UTC m=+1215.214242417" lastFinishedPulling="2025-12-02 10:39:23.245513621 +0000 UTC m=+1216.575652491" observedRunningTime="2025-12-02 10:39:25.419580969 +0000 UTC m=+1218.749719839" watchObservedRunningTime="2025-12-02 10:39:25.422177784 +0000 UTC m=+1218.752316644" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.036952 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.233700 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-scripts\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.233766 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.233832 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data-custom\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.233887 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5985d57a-3142-4a00-840e-060b8bba4e70-logs\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.233963 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-combined-ca-bundle\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.234021 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnkzx\" (UniqueName: \"kubernetes.io/projected/5985d57a-3142-4a00-840e-060b8bba4e70-kube-api-access-jnkzx\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.234085 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5985d57a-3142-4a00-840e-060b8bba4e70-etc-machine-id\") pod \"5985d57a-3142-4a00-840e-060b8bba4e70\" (UID: \"5985d57a-3142-4a00-840e-060b8bba4e70\") " Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.234928 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5985d57a-3142-4a00-840e-060b8bba4e70-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.234969 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5985d57a-3142-4a00-840e-060b8bba4e70-logs" (OuterVolumeSpecName: "logs") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.238467 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.241688 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5985d57a-3142-4a00-840e-060b8bba4e70-kube-api-access-jnkzx" (OuterVolumeSpecName: "kube-api-access-jnkzx") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "kube-api-access-jnkzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.242250 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-scripts" (OuterVolumeSpecName: "scripts") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.258541 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.263621 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.323100 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data" (OuterVolumeSpecName: "config-data") pod "5985d57a-3142-4a00-840e-060b8bba4e70" (UID: "5985d57a-3142-4a00-840e-060b8bba4e70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.336727 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.336963 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.337058 4679 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.337137 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5985d57a-3142-4a00-840e-060b8bba4e70-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.337197 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5985d57a-3142-4a00-840e-060b8bba4e70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.337252 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnkzx\" (UniqueName: \"kubernetes.io/projected/5985d57a-3142-4a00-840e-060b8bba4e70-kube-api-access-jnkzx\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.337316 4679 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5985d57a-3142-4a00-840e-060b8bba4e70-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.396469 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerStarted","Data":"d6956a4f4f3d5e2f0aac5215de4a6bb8e5102fec5f0ab6cad1d61e9197a598bf"} Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.402964 4679 generic.go:334] "Generic (PLEG): container finished" podID="5985d57a-3142-4a00-840e-060b8bba4e70" containerID="f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c" exitCode=0 Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.403000 4679 generic.go:334] "Generic (PLEG): container finished" podID="5985d57a-3142-4a00-840e-060b8bba4e70" containerID="5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae" exitCode=143 Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.404440 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5985d57a-3142-4a00-840e-060b8bba4e70","Type":"ContainerDied","Data":"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c"} Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.404619 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5985d57a-3142-4a00-840e-060b8bba4e70","Type":"ContainerDied","Data":"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae"} Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.404635 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5985d57a-3142-4a00-840e-060b8bba4e70","Type":"ContainerDied","Data":"0490e016c39c405333e564d28772a47c9530bfccf259d219c6fb77aca413735a"} Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.404788 4679 scope.go:117] "RemoveContainer" containerID="f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.405254 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.451963 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.458695 4679 scope.go:117] "RemoveContainer" containerID="5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.493958 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.503500 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:26 crc kubenswrapper[4679]: E1202 10:39:26.504500 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.505058 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api" Dec 02 10:39:26 crc kubenswrapper[4679]: E1202 10:39:26.505202 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api-log" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.505344 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api-log" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.505835 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.505959 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" containerName="cinder-api-log" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.508244 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.511898 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.518799 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.519058 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.519154 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.555965 4679 scope.go:117] "RemoveContainer" containerID="f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c" Dec 02 10:39:26 crc kubenswrapper[4679]: E1202 10:39:26.556444 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c\": container with ID starting with f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c not found: ID does not exist" containerID="f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.556482 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c"} err="failed to get container status \"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c\": rpc error: code = NotFound desc = could not find container \"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c\": container with ID starting with f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c not found: ID does not exist" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.556520 4679 scope.go:117] "RemoveContainer" containerID="5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae" Dec 02 10:39:26 crc kubenswrapper[4679]: E1202 10:39:26.557090 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae\": container with ID starting with 5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae not found: ID does not exist" containerID="5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.557146 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae"} err="failed to get container status \"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae\": rpc error: code = NotFound desc = could not find container \"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae\": container with ID starting with 5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae not found: ID does not exist" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.557190 4679 scope.go:117] "RemoveContainer" containerID="f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.557623 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c"} err="failed to get container status \"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c\": rpc error: code = NotFound desc = could not find container \"f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c\": container with ID starting with f87b2ef9ed80f869d384f31967a18a6e60ff58871f7658a0b34ac2b0df35bf0c not found: ID does not exist" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.557660 4679 scope.go:117] "RemoveContainer" containerID="5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.557875 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae"} err="failed to get container status \"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae\": rpc error: code = NotFound desc = could not find container \"5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae\": container with ID starting with 5c25b92ab208206863d7912d50848117652f9be67a1d4655dc67c9650804feae not found: ID does not exist" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647403 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8a8618-590b-498d-ae0f-ab9f9f83b054-logs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647462 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647493 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-config-data\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647527 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkzbk\" (UniqueName: \"kubernetes.io/projected/5f8a8618-590b-498d-ae0f-ab9f9f83b054-kube-api-access-rkzbk\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647546 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f8a8618-590b-498d-ae0f-ab9f9f83b054-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647804 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.647968 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-scripts\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.648035 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.648085 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-config-data-custom\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.749650 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkzbk\" (UniqueName: \"kubernetes.io/projected/5f8a8618-590b-498d-ae0f-ab9f9f83b054-kube-api-access-rkzbk\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.751107 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f8a8618-590b-498d-ae0f-ab9f9f83b054-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.751378 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f8a8618-590b-498d-ae0f-ab9f9f83b054-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.751692 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.751943 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-scripts\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.752116 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.752361 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-config-data-custom\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.753726 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8a8618-590b-498d-ae0f-ab9f9f83b054-logs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.753967 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.754157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-config-data\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.755273 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f8a8618-590b-498d-ae0f-ab9f9f83b054-logs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.760229 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-config-data\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.761104 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.762136 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.764633 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-scripts\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.765784 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.766167 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f8a8618-590b-498d-ae0f-ab9f9f83b054-config-data-custom\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.778876 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkzbk\" (UniqueName: \"kubernetes.io/projected/5f8a8618-590b-498d-ae0f-ab9f9f83b054-kube-api-access-rkzbk\") pod \"cinder-api-0\" (UID: \"5f8a8618-590b-498d-ae0f-ab9f9f83b054\") " pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.868901 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 10:39:26 crc kubenswrapper[4679]: I1202 10:39:26.929746 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5985d57a-3142-4a00-840e-060b8bba4e70" path="/var/lib/kubelet/pods/5985d57a-3142-4a00-840e-060b8bba4e70/volumes" Dec 02 10:39:27 crc kubenswrapper[4679]: I1202 10:39:27.333191 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 10:39:27 crc kubenswrapper[4679]: I1202 10:39:27.768780 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:39:27 crc kubenswrapper[4679]: I1202 10:39:27.818494 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerStarted","Data":"14ab5f1bb9777a5f1133e5ff3a96419992da3f4c5705aa8c647232d8561b2857"} Dec 02 10:39:27 crc kubenswrapper[4679]: I1202 10:39:27.821243 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5f8a8618-590b-498d-ae0f-ab9f9f83b054","Type":"ContainerStarted","Data":"ca4c21aa35c983b2a10fc04036ae69c22b213709cfef7b909561935f6551300e"} Dec 02 10:39:27 crc kubenswrapper[4679]: I1202 10:39:27.912951 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:39:28 crc kubenswrapper[4679]: I1202 10:39:28.831739 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerStarted","Data":"9d2e16681d29edeb6ff57404bb1449758523eb21ad126002b7c7adab173846a8"} Dec 02 10:39:28 crc kubenswrapper[4679]: I1202 10:39:28.837028 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5f8a8618-590b-498d-ae0f-ab9f9f83b054","Type":"ContainerStarted","Data":"375dc0dd19d59c4e9c84f00aef70fa713c178f3c8c2b621f1a6463bbc646c267"} Dec 02 10:39:28 crc kubenswrapper[4679]: I1202 10:39:28.837075 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5f8a8618-590b-498d-ae0f-ab9f9f83b054","Type":"ContainerStarted","Data":"26df02c24969b9cf9f7c6732409709cd02e51b06e2fff7cd133f6d2a258b7990"} Dec 02 10:39:28 crc kubenswrapper[4679]: I1202 10:39:28.837180 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 02 10:39:28 crc kubenswrapper[4679]: I1202 10:39:28.861866 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.861843183 podStartE2EDuration="2.861843183s" podCreationTimestamp="2025-12-02 10:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:28.852281227 +0000 UTC m=+1222.182420087" watchObservedRunningTime="2025-12-02 10:39:28.861843183 +0000 UTC m=+1222.191982033" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.304518 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:45970->10.217.0.158:9311: read: connection reset by peer" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.304579 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85c6568ff4-fjmqd" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:45978->10.217.0.158:9311: read: connection reset by peer" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.304695 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.498117 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.821092 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6fd99d7d5b-sb96h" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.834976 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.869923 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerStarted","Data":"2c80ad6582afd470e586c5bac30b511d88d607d7c66e3fd8ad9b2c847e938c5a"} Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.870856 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.885106 4679 generic.go:334] "Generic (PLEG): container finished" podID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerID="1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565" exitCode=0 Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.885575 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85c6568ff4-fjmqd" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.886154 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85c6568ff4-fjmqd" event={"ID":"ed234cbf-40c8-4856-bad3-6bf38c7ff610","Type":"ContainerDied","Data":"1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565"} Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.886239 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85c6568ff4-fjmqd" event={"ID":"ed234cbf-40c8-4856-bad3-6bf38c7ff610","Type":"ContainerDied","Data":"a890f66ad3f306f4d21f9097172fe9cb20dc9cf6d1c1064aad408db869c04fe9"} Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.886855 4679 scope.go:117] "RemoveContainer" containerID="1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.917508 4679 scope.go:117] "RemoveContainer" containerID="e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883" Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.961557 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bd7d57448-84dtd"] Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.961877 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bd7d57448-84dtd" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon-log" containerID="cri-o://d87f86cc90eb58eaa5a8aa14c2723787b723ae1ed4cfb00f1c88c0c427250f59" gracePeriod=30 Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.962014 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bd7d57448-84dtd" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" containerID="cri-o://dbbabf0a29747ebfd44364b9b3f7b60a84d413aee3ffe565fee42f380f0ae98f" gracePeriod=30 Dec 02 10:39:29 crc kubenswrapper[4679]: I1202 10:39:29.993835 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.189884903 podStartE2EDuration="6.993814609s" podCreationTimestamp="2025-12-02 10:39:23 +0000 UTC" firstStartedPulling="2025-12-02 10:39:24.635066767 +0000 UTC m=+1217.965205627" lastFinishedPulling="2025-12-02 10:39:29.438996473 +0000 UTC m=+1222.769135333" observedRunningTime="2025-12-02 10:39:29.960651192 +0000 UTC m=+1223.290790052" watchObservedRunningTime="2025-12-02 10:39:29.993814609 +0000 UTC m=+1223.323953469" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.002712 4679 scope.go:117] "RemoveContainer" containerID="1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565" Dec 02 10:39:30 crc kubenswrapper[4679]: E1202 10:39:30.003619 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565\": container with ID starting with 1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565 not found: ID does not exist" containerID="1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.003666 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565"} err="failed to get container status \"1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565\": rpc error: code = NotFound desc = could not find container \"1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565\": container with ID starting with 1e2675523b8b21a8ad6b97721a40c11eddd2d2626b145bcc976d0ec648fd9565 not found: ID does not exist" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.003693 4679 scope.go:117] "RemoveContainer" containerID="e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883" Dec 02 10:39:30 crc kubenswrapper[4679]: E1202 10:39:30.004224 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883\": container with ID starting with e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883 not found: ID does not exist" containerID="e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.004252 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883"} err="failed to get container status \"e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883\": rpc error: code = NotFound desc = could not find container \"e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883\": container with ID starting with e3d589b22cfe8a775783045c84e2f653e8ec159899398a6f9453ac68fa43b883 not found: ID does not exist" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.026889 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-combined-ca-bundle\") pod \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.026943 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data\") pod \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.027016 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed234cbf-40c8-4856-bad3-6bf38c7ff610-logs\") pod \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.027072 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data-custom\") pod \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.027141 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ggv6\" (UniqueName: \"kubernetes.io/projected/ed234cbf-40c8-4856-bad3-6bf38c7ff610-kube-api-access-9ggv6\") pod \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\" (UID: \"ed234cbf-40c8-4856-bad3-6bf38c7ff610\") " Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.027703 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed234cbf-40c8-4856-bad3-6bf38c7ff610-logs" (OuterVolumeSpecName: "logs") pod "ed234cbf-40c8-4856-bad3-6bf38c7ff610" (UID: "ed234cbf-40c8-4856-bad3-6bf38c7ff610"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.028057 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed234cbf-40c8-4856-bad3-6bf38c7ff610-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.033004 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ed234cbf-40c8-4856-bad3-6bf38c7ff610" (UID: "ed234cbf-40c8-4856-bad3-6bf38c7ff610"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.033039 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed234cbf-40c8-4856-bad3-6bf38c7ff610-kube-api-access-9ggv6" (OuterVolumeSpecName: "kube-api-access-9ggv6") pod "ed234cbf-40c8-4856-bad3-6bf38c7ff610" (UID: "ed234cbf-40c8-4856-bad3-6bf38c7ff610"). InnerVolumeSpecName "kube-api-access-9ggv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.051855 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed234cbf-40c8-4856-bad3-6bf38c7ff610" (UID: "ed234cbf-40c8-4856-bad3-6bf38c7ff610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.085863 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data" (OuterVolumeSpecName: "config-data") pod "ed234cbf-40c8-4856-bad3-6bf38c7ff610" (UID: "ed234cbf-40c8-4856-bad3-6bf38c7ff610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.121600 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6dbdc9c48f-dtqv6" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.131174 4679 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.131201 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ggv6\" (UniqueName: \"kubernetes.io/projected/ed234cbf-40c8-4856-bad3-6bf38c7ff610-kube-api-access-9ggv6\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.131211 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.131220 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed234cbf-40c8-4856-bad3-6bf38c7ff610-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.194079 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-559ffdfb7b-75pb4"] Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.194594 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-559ffdfb7b-75pb4" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-api" containerID="cri-o://808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155" gracePeriod=30 Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.194718 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-559ffdfb7b-75pb4" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-httpd" containerID="cri-o://7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67" gracePeriod=30 Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.243677 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85c6568ff4-fjmqd"] Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.251850 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-85c6568ff4-fjmqd"] Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.895261 4679 generic.go:334] "Generic (PLEG): container finished" podID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerID="7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67" exitCode=0 Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.895344 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-559ffdfb7b-75pb4" event={"ID":"3d11f0a0-986a-4012-bea2-a6c5b325cb7a","Type":"ContainerDied","Data":"7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67"} Dec 02 10:39:30 crc kubenswrapper[4679]: I1202 10:39:30.921873 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" path="/var/lib/kubelet/pods/ed234cbf-40c8-4856-bad3-6bf38c7ff610/volumes" Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.424575 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.468272 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.488561 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vggwm"] Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.488870 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="dnsmasq-dns" containerID="cri-o://40bd0ffa15a8f1c691c47a4c9d9e744e3a368de748fe3021427a268baebdb592" gracePeriod=10 Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.517831 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.906435 4679 generic.go:334] "Generic (PLEG): container finished" podID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerID="40bd0ffa15a8f1c691c47a4c9d9e744e3a368de748fe3021427a268baebdb592" exitCode=0 Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.906634 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="cinder-scheduler" containerID="cri-o://e16ba62420535f6092ea8d7a69351e248278777a86ae35326ef77b227a7b17a6" gracePeriod=30 Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.906958 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" event={"ID":"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce","Type":"ContainerDied","Data":"40bd0ffa15a8f1c691c47a4c9d9e744e3a368de748fe3021427a268baebdb592"} Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.906985 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" event={"ID":"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce","Type":"ContainerDied","Data":"863eece2e504fb8187ab2245295d21507dc65a46b1c4672133d7dbd7ed39c958"} Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.906995 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="863eece2e504fb8187ab2245295d21507dc65a46b1c4672133d7dbd7ed39c958" Dec 02 10:39:31 crc kubenswrapper[4679]: I1202 10:39:31.907168 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="probe" containerID="cri-o://472f686773c5600ae7db88388b27ec92d15c5cddfdcfc856dfeb8c59ee4936f6" gracePeriod=30 Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.012179 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.070799 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-sb\") pod \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.070911 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-nb\") pod \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.070963 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7c82\" (UniqueName: \"kubernetes.io/projected/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-kube-api-access-m7c82\") pod \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.071012 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-swift-storage-0\") pod \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.071054 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-svc\") pod \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.071100 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-config\") pod \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\" (UID: \"bb2d39c2-e2dd-44c7-aaab-094e133aa6ce\") " Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.104522 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-kube-api-access-m7c82" (OuterVolumeSpecName: "kube-api-access-m7c82") pod "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" (UID: "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce"). InnerVolumeSpecName "kube-api-access-m7c82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.138605 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-config" (OuterVolumeSpecName: "config") pod "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" (UID: "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.145548 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" (UID: "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.145560 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" (UID: "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.153013 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" (UID: "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.156598 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" (UID: "bb2d39c2-e2dd-44c7-aaab-094e133aa6ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.173571 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.173606 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.173617 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7c82\" (UniqueName: \"kubernetes.io/projected/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-kube-api-access-m7c82\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.173631 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.173641 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.173651 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.917832 4679 generic.go:334] "Generic (PLEG): container finished" podID="79a91754-89a4-44e6-8f0b-e511785302ad" containerID="472f686773c5600ae7db88388b27ec92d15c5cddfdcfc856dfeb8c59ee4936f6" exitCode=0 Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.918270 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.919811 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79a91754-89a4-44e6-8f0b-e511785302ad","Type":"ContainerDied","Data":"472f686773c5600ae7db88388b27ec92d15c5cddfdcfc856dfeb8c59ee4936f6"} Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.959661 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vggwm"] Dec 02 10:39:32 crc kubenswrapper[4679]: I1202 10:39:32.968952 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vggwm"] Dec 02 10:39:33 crc kubenswrapper[4679]: I1202 10:39:33.928715 4679 generic.go:334] "Generic (PLEG): container finished" podID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerID="dbbabf0a29747ebfd44364b9b3f7b60a84d413aee3ffe565fee42f380f0ae98f" exitCode=0 Dec 02 10:39:33 crc kubenswrapper[4679]: I1202 10:39:33.928748 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bd7d57448-84dtd" event={"ID":"ded5eaca-d45b-46bf-a770-50d909f0003e","Type":"ContainerDied","Data":"dbbabf0a29747ebfd44364b9b3f7b60a84d413aee3ffe565fee42f380f0ae98f"} Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.596288 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.610708 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r62l\" (UniqueName: \"kubernetes.io/projected/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-kube-api-access-7r62l\") pod \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.610790 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-ovndb-tls-certs\") pod \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.610822 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-combined-ca-bundle\") pod \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.610955 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-config\") pod \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.610998 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-httpd-config\") pod \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\" (UID: \"3d11f0a0-986a-4012-bea2-a6c5b325cb7a\") " Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.617431 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3d11f0a0-986a-4012-bea2-a6c5b325cb7a" (UID: "3d11f0a0-986a-4012-bea2-a6c5b325cb7a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.619448 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-kube-api-access-7r62l" (OuterVolumeSpecName: "kube-api-access-7r62l") pod "3d11f0a0-986a-4012-bea2-a6c5b325cb7a" (UID: "3d11f0a0-986a-4012-bea2-a6c5b325cb7a"). InnerVolumeSpecName "kube-api-access-7r62l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.679945 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-config" (OuterVolumeSpecName: "config") pod "3d11f0a0-986a-4012-bea2-a6c5b325cb7a" (UID: "3d11f0a0-986a-4012-bea2-a6c5b325cb7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.688692 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d11f0a0-986a-4012-bea2-a6c5b325cb7a" (UID: "3d11f0a0-986a-4012-bea2-a6c5b325cb7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.705659 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "3d11f0a0-986a-4012-bea2-a6c5b325cb7a" (UID: "3d11f0a0-986a-4012-bea2-a6c5b325cb7a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.713727 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.713775 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.713788 4679 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.713801 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r62l\" (UniqueName: \"kubernetes.io/projected/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-kube-api-access-7r62l\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.713815 4679 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d11f0a0-986a-4012-bea2-a6c5b325cb7a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.918981 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" path="/var/lib/kubelet/pods/bb2d39c2-e2dd-44c7-aaab-094e133aa6ce/volumes" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.938558 4679 generic.go:334] "Generic (PLEG): container finished" podID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerID="808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155" exitCode=0 Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.938613 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-559ffdfb7b-75pb4" event={"ID":"3d11f0a0-986a-4012-bea2-a6c5b325cb7a","Type":"ContainerDied","Data":"808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155"} Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.938623 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-559ffdfb7b-75pb4" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.938653 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-559ffdfb7b-75pb4" event={"ID":"3d11f0a0-986a-4012-bea2-a6c5b325cb7a","Type":"ContainerDied","Data":"b3cfde8b41fe8c0ebd88b94cae96b301b8f7b41d0b3b24851e19004ec57e8bdc"} Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.938674 4679 scope.go:117] "RemoveContainer" containerID="7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67" Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.961356 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-559ffdfb7b-75pb4"] Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.968028 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-559ffdfb7b-75pb4"] Dec 02 10:39:34 crc kubenswrapper[4679]: I1202 10:39:34.969201 4679 scope.go:117] "RemoveContainer" containerID="808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155" Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.000362 4679 scope.go:117] "RemoveContainer" containerID="7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67" Dec 02 10:39:35 crc kubenswrapper[4679]: E1202 10:39:35.001027 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67\": container with ID starting with 7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67 not found: ID does not exist" containerID="7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67" Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.001057 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67"} err="failed to get container status \"7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67\": rpc error: code = NotFound desc = could not find container \"7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67\": container with ID starting with 7ab38ac0e8d4af6897c61ee842bc3d7d5fcf2ecd3ba5df5583e41a030f265a67 not found: ID does not exist" Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.001080 4679 scope.go:117] "RemoveContainer" containerID="808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155" Dec 02 10:39:35 crc kubenswrapper[4679]: E1202 10:39:35.001519 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155\": container with ID starting with 808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155 not found: ID does not exist" containerID="808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155" Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.001569 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155"} err="failed to get container status \"808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155\": rpc error: code = NotFound desc = could not find container \"808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155\": container with ID starting with 808afe7dbe487e553579752e8fcd3335c33d7e668468868f5907848fc1f3a155 not found: ID does not exist" Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.298983 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6bd7d57448-84dtd" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.953489 4679 generic.go:334] "Generic (PLEG): container finished" podID="79a91754-89a4-44e6-8f0b-e511785302ad" containerID="e16ba62420535f6092ea8d7a69351e248278777a86ae35326ef77b227a7b17a6" exitCode=0 Dec 02 10:39:35 crc kubenswrapper[4679]: I1202 10:39:35.953822 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79a91754-89a4-44e6-8f0b-e511785302ad","Type":"ContainerDied","Data":"e16ba62420535f6092ea8d7a69351e248278777a86ae35326ef77b227a7b17a6"} Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.317710 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.346992 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a91754-89a4-44e6-8f0b-e511785302ad-etc-machine-id\") pod \"79a91754-89a4-44e6-8f0b-e511785302ad\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347126 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79a91754-89a4-44e6-8f0b-e511785302ad-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "79a91754-89a4-44e6-8f0b-e511785302ad" (UID: "79a91754-89a4-44e6-8f0b-e511785302ad"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347166 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data\") pod \"79a91754-89a4-44e6-8f0b-e511785302ad\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347213 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zbfl\" (UniqueName: \"kubernetes.io/projected/79a91754-89a4-44e6-8f0b-e511785302ad-kube-api-access-9zbfl\") pod \"79a91754-89a4-44e6-8f0b-e511785302ad\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347285 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-scripts\") pod \"79a91754-89a4-44e6-8f0b-e511785302ad\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347346 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-combined-ca-bundle\") pod \"79a91754-89a4-44e6-8f0b-e511785302ad\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347375 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data-custom\") pod \"79a91754-89a4-44e6-8f0b-e511785302ad\" (UID: \"79a91754-89a4-44e6-8f0b-e511785302ad\") " Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.347908 4679 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79a91754-89a4-44e6-8f0b-e511785302ad-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.356388 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-scripts" (OuterVolumeSpecName: "scripts") pod "79a91754-89a4-44e6-8f0b-e511785302ad" (UID: "79a91754-89a4-44e6-8f0b-e511785302ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.357055 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "79a91754-89a4-44e6-8f0b-e511785302ad" (UID: "79a91754-89a4-44e6-8f0b-e511785302ad"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.373497 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a91754-89a4-44e6-8f0b-e511785302ad-kube-api-access-9zbfl" (OuterVolumeSpecName: "kube-api-access-9zbfl") pod "79a91754-89a4-44e6-8f0b-e511785302ad" (UID: "79a91754-89a4-44e6-8f0b-e511785302ad"). InnerVolumeSpecName "kube-api-access-9zbfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.449178 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.449211 4679 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.449224 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zbfl\" (UniqueName: \"kubernetes.io/projected/79a91754-89a4-44e6-8f0b-e511785302ad-kube-api-access-9zbfl\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.457524 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79a91754-89a4-44e6-8f0b-e511785302ad" (UID: "79a91754-89a4-44e6-8f0b-e511785302ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.515937 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data" (OuterVolumeSpecName: "config-data") pod "79a91754-89a4-44e6-8f0b-e511785302ad" (UID: "79a91754-89a4-44e6-8f0b-e511785302ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.549825 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.549852 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a91754-89a4-44e6-8f0b-e511785302ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.836021 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75c8ddd69c-vggwm" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.156:5353: i/o timeout" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.920750 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" path="/var/lib/kubelet/pods/3d11f0a0-986a-4012-bea2-a6c5b325cb7a/volumes" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.966414 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79a91754-89a4-44e6-8f0b-e511785302ad","Type":"ContainerDied","Data":"404330e1aa3b7e0399731a0e8d38288be0a26b32d057eca2dff38d70d4abfc48"} Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.966465 4679 scope.go:117] "RemoveContainer" containerID="472f686773c5600ae7db88388b27ec92d15c5cddfdcfc856dfeb8c59ee4936f6" Dec 02 10:39:36 crc kubenswrapper[4679]: I1202 10:39:36.966520 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.001505 4679 scope.go:117] "RemoveContainer" containerID="e16ba62420535f6092ea8d7a69351e248278777a86ae35326ef77b227a7b17a6" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.003549 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.029846 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043272 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043714 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="init" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043730 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="init" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043743 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="dnsmasq-dns" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043751 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="dnsmasq-dns" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043764 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="probe" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043773 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="probe" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043780 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-api" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043787 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-api" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043810 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043816 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043834 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-httpd" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043840 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-httpd" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043849 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api-log" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043855 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api-log" Dec 02 10:39:37 crc kubenswrapper[4679]: E1202 10:39:37.043862 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="cinder-scheduler" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.043867 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="cinder-scheduler" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044037 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044049 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed234cbf-40c8-4856-bad3-6bf38c7ff610" containerName="barbican-api-log" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044060 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="cinder-scheduler" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044073 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" containerName="probe" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044087 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-httpd" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044094 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d11f0a0-986a-4012-bea2-a6c5b325cb7a" containerName="neutron-api" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.044102 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb2d39c2-e2dd-44c7-aaab-094e133aa6ce" containerName="dnsmasq-dns" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.045007 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.048853 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.068601 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vhxn\" (UniqueName: \"kubernetes.io/projected/247813fb-2103-4014-b0a3-c5be6310a181-kube-api-access-4vhxn\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.068660 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.068684 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.068706 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-config-data\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.068800 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-scripts\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.068832 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/247813fb-2103-4014-b0a3-c5be6310a181-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.078644 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170253 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-scripts\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170580 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/247813fb-2103-4014-b0a3-c5be6310a181-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170626 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vhxn\" (UniqueName: \"kubernetes.io/projected/247813fb-2103-4014-b0a3-c5be6310a181-kube-api-access-4vhxn\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170662 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170682 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170705 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-config-data\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.170741 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/247813fb-2103-4014-b0a3-c5be6310a181-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.175815 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.177029 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-config-data\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.178908 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.186620 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/247813fb-2103-4014-b0a3-c5be6310a181-scripts\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.190788 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vhxn\" (UniqueName: \"kubernetes.io/projected/247813fb-2103-4014-b0a3-c5be6310a181-kube-api-access-4vhxn\") pod \"cinder-scheduler-0\" (UID: \"247813fb-2103-4014-b0a3-c5be6310a181\") " pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.376607 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 10:39:37 crc kubenswrapper[4679]: I1202 10:39:37.869032 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 10:39:38 crc kubenswrapper[4679]: I1202 10:39:38.057133 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"247813fb-2103-4014-b0a3-c5be6310a181","Type":"ContainerStarted","Data":"751aad3c6c02f9f3061ef6e2de797e4215fff28b0602580cdd62e29a55f4e1c4"} Dec 02 10:39:38 crc kubenswrapper[4679]: I1202 10:39:38.643906 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6558c68964-dhdww" Dec 02 10:39:38 crc kubenswrapper[4679]: I1202 10:39:38.924891 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a91754-89a4-44e6-8f0b-e511785302ad" path="/var/lib/kubelet/pods/79a91754-89a4-44e6-8f0b-e511785302ad/volumes" Dec 02 10:39:39 crc kubenswrapper[4679]: I1202 10:39:39.018951 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 02 10:39:39 crc kubenswrapper[4679]: I1202 10:39:39.080443 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"247813fb-2103-4014-b0a3-c5be6310a181","Type":"ContainerStarted","Data":"f4ae00ce8a2ec2e11c4268f63e10cda99ea1224a29e98263b6bc1143e501a6ce"} Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.033641 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.035087 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.037021 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.037098 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-dhkkg" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.040324 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.057379 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.108716 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"247813fb-2103-4014-b0a3-c5be6310a181","Type":"ContainerStarted","Data":"bbc6439789e8105ac996a385e805105b9cf2746b5aad384189557aa1b2d646be"} Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.133718 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.133698148 podStartE2EDuration="3.133698148s" podCreationTimestamp="2025-12-02 10:39:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:40.127462888 +0000 UTC m=+1233.457601768" watchObservedRunningTime="2025-12-02 10:39:40.133698148 +0000 UTC m=+1233.463837008" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.138031 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d970b7-84af-4667-8b4e-6b2d4b800923-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.138162 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0d970b7-84af-4667-8b4e-6b2d4b800923-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.138271 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0d970b7-84af-4667-8b4e-6b2d4b800923-openstack-config\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.138296 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7czb4\" (UniqueName: \"kubernetes.io/projected/c0d970b7-84af-4667-8b4e-6b2d4b800923-kube-api-access-7czb4\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.240566 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d970b7-84af-4667-8b4e-6b2d4b800923-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.240734 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0d970b7-84af-4667-8b4e-6b2d4b800923-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.240814 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0d970b7-84af-4667-8b4e-6b2d4b800923-openstack-config\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.240841 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7czb4\" (UniqueName: \"kubernetes.io/projected/c0d970b7-84af-4667-8b4e-6b2d4b800923-kube-api-access-7czb4\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.242768 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0d970b7-84af-4667-8b4e-6b2d4b800923-openstack-config\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.250511 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0d970b7-84af-4667-8b4e-6b2d4b800923-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.265109 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d970b7-84af-4667-8b4e-6b2d4b800923-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.266877 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7czb4\" (UniqueName: \"kubernetes.io/projected/c0d970b7-84af-4667-8b4e-6b2d4b800923-kube-api-access-7czb4\") pod \"openstackclient\" (UID: \"c0d970b7-84af-4667-8b4e-6b2d4b800923\") " pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.357037 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 10:39:40 crc kubenswrapper[4679]: I1202 10:39:40.930379 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 10:39:41 crc kubenswrapper[4679]: I1202 10:39:41.120331 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c0d970b7-84af-4667-8b4e-6b2d4b800923","Type":"ContainerStarted","Data":"250cf8b18fdcd85544bfcf30a556569f473310f7a5a14cf9cedc65d89e6f4f93"} Dec 02 10:39:42 crc kubenswrapper[4679]: I1202 10:39:42.376808 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.479974 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-8d8658b7-4bkfp"] Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.481955 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.484768 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.485120 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.485246 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.519755 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8d8658b7-4bkfp"] Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611092 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-internal-tls-certs\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611149 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-public-tls-certs\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611171 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8d434783-1fe8-4ba4-8195-282fa213066d-etc-swift\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611193 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr842\" (UniqueName: \"kubernetes.io/projected/8d434783-1fe8-4ba4-8195-282fa213066d-kube-api-access-hr842\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611221 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-config-data\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611242 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-combined-ca-bundle\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611285 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d434783-1fe8-4ba4-8195-282fa213066d-run-httpd\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.611315 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d434783-1fe8-4ba4-8195-282fa213066d-log-httpd\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713049 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d434783-1fe8-4ba4-8195-282fa213066d-run-httpd\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713097 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d434783-1fe8-4ba4-8195-282fa213066d-log-httpd\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713226 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-internal-tls-certs\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713271 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-public-tls-certs\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713363 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8d434783-1fe8-4ba4-8195-282fa213066d-etc-swift\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713400 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr842\" (UniqueName: \"kubernetes.io/projected/8d434783-1fe8-4ba4-8195-282fa213066d-kube-api-access-hr842\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713441 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-config-data\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713472 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-combined-ca-bundle\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713529 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d434783-1fe8-4ba4-8195-282fa213066d-run-httpd\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.713603 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d434783-1fe8-4ba4-8195-282fa213066d-log-httpd\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.721285 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-config-data\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.728497 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-public-tls-certs\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.730337 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-combined-ca-bundle\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.732022 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr842\" (UniqueName: \"kubernetes.io/projected/8d434783-1fe8-4ba4-8195-282fa213066d-kube-api-access-hr842\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.732461 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8d434783-1fe8-4ba4-8195-282fa213066d-etc-swift\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.743274 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d434783-1fe8-4ba4-8195-282fa213066d-internal-tls-certs\") pod \"swift-proxy-8d8658b7-4bkfp\" (UID: \"8d434783-1fe8-4ba4-8195-282fa213066d\") " pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:43 crc kubenswrapper[4679]: I1202 10:39:43.818753 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:44 crc kubenswrapper[4679]: I1202 10:39:44.394973 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8d8658b7-4bkfp"] Dec 02 10:39:44 crc kubenswrapper[4679]: I1202 10:39:44.567400 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:44 crc kubenswrapper[4679]: I1202 10:39:44.567662 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-log" containerID="cri-o://b87d66c33269519376e092d1b0bf22a0441cc598188d0a792f9a159942fc33bc" gracePeriod=30 Dec 02 10:39:44 crc kubenswrapper[4679]: I1202 10:39:44.567776 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-httpd" containerID="cri-o://b36b75676ab9c393bc54e903d245097061243d086034ab24a6f56ca3f1bebd72" gracePeriod=30 Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.162898 4679 generic.go:334] "Generic (PLEG): container finished" podID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerID="b87d66c33269519376e092d1b0bf22a0441cc598188d0a792f9a159942fc33bc" exitCode=143 Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.162981 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8fae901a-a069-4931-adad-b71b0cd4df8e","Type":"ContainerDied","Data":"b87d66c33269519376e092d1b0bf22a0441cc598188d0a792f9a159942fc33bc"} Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.165819 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8d8658b7-4bkfp" event={"ID":"8d434783-1fe8-4ba4-8195-282fa213066d","Type":"ContainerStarted","Data":"50891e0e202d62de50163d4fd0361f6b15de2e8077e9ba4ac12bcbc5fe168932"} Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.165866 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8d8658b7-4bkfp" event={"ID":"8d434783-1fe8-4ba4-8195-282fa213066d","Type":"ContainerStarted","Data":"19879539bd55e905cfcbc14be73f332018c2e4b83252849bd8b9cd4c2c51ad41"} Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.165877 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8d8658b7-4bkfp" event={"ID":"8d434783-1fe8-4ba4-8195-282fa213066d","Type":"ContainerStarted","Data":"6e11f31361935c4ae448e8d2cf4649e26f547fdc48323a8e4dfe4b95b02016be"} Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.166041 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.192789 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-8d8658b7-4bkfp" podStartSLOduration=2.192768464 podStartE2EDuration="2.192768464s" podCreationTimestamp="2025-12-02 10:39:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:45.188748788 +0000 UTC m=+1238.518887658" watchObservedRunningTime="2025-12-02 10:39:45.192768464 +0000 UTC m=+1238.522907324" Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.285653 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.285923 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-central-agent" containerID="cri-o://d6956a4f4f3d5e2f0aac5215de4a6bb8e5102fec5f0ab6cad1d61e9197a598bf" gracePeriod=30 Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.286067 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="proxy-httpd" containerID="cri-o://2c80ad6582afd470e586c5bac30b511d88d607d7c66e3fd8ad9b2c847e938c5a" gracePeriod=30 Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.286101 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="sg-core" containerID="cri-o://9d2e16681d29edeb6ff57404bb1449758523eb21ad126002b7c7adab173846a8" gracePeriod=30 Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.286134 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-notification-agent" containerID="cri-o://14ab5f1bb9777a5f1133e5ff3a96419992da3f4c5705aa8c647232d8561b2857" gracePeriod=30 Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.295450 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.163:3000/\": EOF" Dec 02 10:39:45 crc kubenswrapper[4679]: I1202 10:39:45.297903 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6bd7d57448-84dtd" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 02 10:39:45 crc kubenswrapper[4679]: E1202 10:39:45.592649 4679 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27679d46_4a68_4372_be27_9a9ceb14ac61.slice/crio-2c80ad6582afd470e586c5bac30b511d88d607d7c66e3fd8ad9b2c847e938c5a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27679d46_4a68_4372_be27_9a9ceb14ac61.slice/crio-conmon-9d2e16681d29edeb6ff57404bb1449758523eb21ad126002b7c7adab173846a8.scope\": RecentStats: unable to find data in memory cache]" Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.193642 4679 generic.go:334] "Generic (PLEG): container finished" podID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerID="2c80ad6582afd470e586c5bac30b511d88d607d7c66e3fd8ad9b2c847e938c5a" exitCode=0 Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.193675 4679 generic.go:334] "Generic (PLEG): container finished" podID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerID="9d2e16681d29edeb6ff57404bb1449758523eb21ad126002b7c7adab173846a8" exitCode=2 Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.193683 4679 generic.go:334] "Generic (PLEG): container finished" podID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerID="d6956a4f4f3d5e2f0aac5215de4a6bb8e5102fec5f0ab6cad1d61e9197a598bf" exitCode=0 Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.194451 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerDied","Data":"2c80ad6582afd470e586c5bac30b511d88d607d7c66e3fd8ad9b2c847e938c5a"} Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.194479 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerDied","Data":"9d2e16681d29edeb6ff57404bb1449758523eb21ad126002b7c7adab173846a8"} Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.194489 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerDied","Data":"d6956a4f4f3d5e2f0aac5215de4a6bb8e5102fec5f0ab6cad1d61e9197a598bf"} Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.194510 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.932900 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:39:46 crc kubenswrapper[4679]: I1202 10:39:46.933269 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:39:47 crc kubenswrapper[4679]: I1202 10:39:47.147424 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:47 crc kubenswrapper[4679]: I1202 10:39:47.147671 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-log" containerID="cri-o://a2c1e359e1d0038891a7043726146cd0875dec70ec23bd403e0a37eec1b35286" gracePeriod=30 Dec 02 10:39:47 crc kubenswrapper[4679]: I1202 10:39:47.147823 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-httpd" containerID="cri-o://8c7f669e05352f9712222cdbe0ad49be5385a980e06c61e53e2a95a934486904" gracePeriod=30 Dec 02 10:39:47 crc kubenswrapper[4679]: I1202 10:39:47.582792 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 02 10:39:48 crc kubenswrapper[4679]: I1202 10:39:48.215960 4679 generic.go:334] "Generic (PLEG): container finished" podID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerID="a2c1e359e1d0038891a7043726146cd0875dec70ec23bd403e0a37eec1b35286" exitCode=143 Dec 02 10:39:48 crc kubenswrapper[4679]: I1202 10:39:48.216150 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5799263-7d7f-4195-b6b2-70d9315e3a3f","Type":"ContainerDied","Data":"a2c1e359e1d0038891a7043726146cd0875dec70ec23bd403e0a37eec1b35286"} Dec 02 10:39:48 crc kubenswrapper[4679]: I1202 10:39:48.219760 4679 generic.go:334] "Generic (PLEG): container finished" podID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerID="b36b75676ab9c393bc54e903d245097061243d086034ab24a6f56ca3f1bebd72" exitCode=0 Dec 02 10:39:48 crc kubenswrapper[4679]: I1202 10:39:48.219806 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8fae901a-a069-4931-adad-b71b0cd4df8e","Type":"ContainerDied","Data":"b36b75676ab9c393bc54e903d245097061243d086034ab24a6f56ca3f1bebd72"} Dec 02 10:39:48 crc kubenswrapper[4679]: I1202 10:39:48.222897 4679 generic.go:334] "Generic (PLEG): container finished" podID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerID="14ab5f1bb9777a5f1133e5ff3a96419992da3f4c5705aa8c647232d8561b2857" exitCode=0 Dec 02 10:39:48 crc kubenswrapper[4679]: I1202 10:39:48.222933 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerDied","Data":"14ab5f1bb9777a5f1133e5ff3a96419992da3f4c5705aa8c647232d8561b2857"} Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.252558 4679 generic.go:334] "Generic (PLEG): container finished" podID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerID="8c7f669e05352f9712222cdbe0ad49be5385a980e06c61e53e2a95a934486904" exitCode=0 Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.253031 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5799263-7d7f-4195-b6b2-70d9315e3a3f","Type":"ContainerDied","Data":"8c7f669e05352f9712222cdbe0ad49be5385a980e06c61e53e2a95a934486904"} Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.409986 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.479061 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555792 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-combined-ca-bundle\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555835 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-httpd-run\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555857 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zz4p\" (UniqueName: \"kubernetes.io/projected/8fae901a-a069-4931-adad-b71b0cd4df8e-kube-api-access-6zz4p\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555881 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-internal-tls-certs\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555912 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-scripts\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555945 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-logs\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.555965 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-config-data\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.556044 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8fae901a-a069-4931-adad-b71b0cd4df8e\" (UID: \"8fae901a-a069-4931-adad-b71b0cd4df8e\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.559328 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.561061 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-logs" (OuterVolumeSpecName: "logs") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.566083 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.566568 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-scripts" (OuterVolumeSpecName: "scripts") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.566133 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fae901a-a069-4931-adad-b71b0cd4df8e-kube-api-access-6zz4p" (OuterVolumeSpecName: "kube-api-access-6zz4p") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "kube-api-access-6zz4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.585626 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.589908 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.617787 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-config-data" (OuterVolumeSpecName: "config-data") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.645881 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8fae901a-a069-4931-adad-b71b0cd4df8e" (UID: "8fae901a-a069-4931-adad-b71b0cd4df8e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.657550 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-combined-ca-bundle\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.657611 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-scripts\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.657849 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtnlb\" (UniqueName: \"kubernetes.io/projected/27679d46-4a68-4372-be27-9a9ceb14ac61-kube-api-access-xtnlb\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.657893 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-log-httpd\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.657963 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-sg-core-conf-yaml\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658029 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-run-httpd\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658047 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-config-data\") pod \"27679d46-4a68-4372-be27-9a9ceb14ac61\" (UID: \"27679d46-4a68-4372-be27-9a9ceb14ac61\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658429 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658454 4679 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658469 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zz4p\" (UniqueName: \"kubernetes.io/projected/8fae901a-a069-4931-adad-b71b0cd4df8e-kube-api-access-6zz4p\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658483 4679 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658495 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658509 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fae901a-a069-4931-adad-b71b0cd4df8e-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658522 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fae901a-a069-4931-adad-b71b0cd4df8e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658549 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.658999 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.659162 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.661508 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-scripts" (OuterVolumeSpecName: "scripts") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.661957 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27679d46-4a68-4372-be27-9a9ceb14ac61-kube-api-access-xtnlb" (OuterVolumeSpecName: "kube-api-access-xtnlb") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "kube-api-access-xtnlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.682039 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.692106 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.729791 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.749530 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-config-data" (OuterVolumeSpecName: "config-data") pod "27679d46-4a68-4372-be27-9a9ceb14ac61" (UID: "27679d46-4a68-4372-be27-9a9ceb14ac61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.759931 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-config-data\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.759994 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-httpd-run\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760045 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760087 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-public-tls-certs\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760120 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-scripts\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760223 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-combined-ca-bundle\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760289 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-logs\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760363 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbvm6\" (UniqueName: \"kubernetes.io/projected/c5799263-7d7f-4195-b6b2-70d9315e3a3f-kube-api-access-sbvm6\") pod \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\" (UID: \"c5799263-7d7f-4195-b6b2-70d9315e3a3f\") " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760695 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760713 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760723 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760732 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760743 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760751 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27679d46-4a68-4372-be27-9a9ceb14ac61-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760759 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtnlb\" (UniqueName: \"kubernetes.io/projected/27679d46-4a68-4372-be27-9a9ceb14ac61-kube-api-access-xtnlb\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.760768 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/27679d46-4a68-4372-be27-9a9ceb14ac61-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.761492 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.762138 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-logs" (OuterVolumeSpecName: "logs") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.763992 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5799263-7d7f-4195-b6b2-70d9315e3a3f-kube-api-access-sbvm6" (OuterVolumeSpecName: "kube-api-access-sbvm6") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "kube-api-access-sbvm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.763993 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.764641 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-scripts" (OuterVolumeSpecName: "scripts") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.785265 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.803135 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-config-data" (OuterVolumeSpecName: "config-data") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.805300 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c5799263-7d7f-4195-b6b2-70d9315e3a3f" (UID: "c5799263-7d7f-4195-b6b2-70d9315e3a3f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863026 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863335 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863461 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbvm6\" (UniqueName: \"kubernetes.io/projected/c5799263-7d7f-4195-b6b2-70d9315e3a3f-kube-api-access-sbvm6\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863590 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863718 4679 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5799263-7d7f-4195-b6b2-70d9315e3a3f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863843 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.863936 4679 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.864033 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5799263-7d7f-4195-b6b2-70d9315e3a3f-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.884457 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 02 10:39:51 crc kubenswrapper[4679]: I1202 10:39:51.966086 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.263426 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"27679d46-4a68-4372-be27-9a9ceb14ac61","Type":"ContainerDied","Data":"5f813632e448a1275ef1a9b5a3d8a288ad450ed6371cadb94fc60058fa6357bb"} Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.263793 4679 scope.go:117] "RemoveContainer" containerID="2c80ad6582afd470e586c5bac30b511d88d607d7c66e3fd8ad9b2c847e938c5a" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.263478 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.266324 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c0d970b7-84af-4667-8b4e-6b2d4b800923","Type":"ContainerStarted","Data":"57cb98f3c2c6935a9a0fbb0bdcda13e04cdaeb3fc6cf71f676044aa8b1683c10"} Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.270056 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5799263-7d7f-4195-b6b2-70d9315e3a3f","Type":"ContainerDied","Data":"869d8d431496fb084a9e88b79149dd26f397905071c2589db2c265362644bb68"} Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.270161 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.275863 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8fae901a-a069-4931-adad-b71b0cd4df8e","Type":"ContainerDied","Data":"91e90640aede6ecb4361843fe8885180cb2b947b377cee9304f7db57be2d6173"} Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.275921 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.292150 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.086642857 podStartE2EDuration="12.292133479s" podCreationTimestamp="2025-12-02 10:39:40 +0000 UTC" firstStartedPulling="2025-12-02 10:39:40.938882956 +0000 UTC m=+1234.269021816" lastFinishedPulling="2025-12-02 10:39:51.144373578 +0000 UTC m=+1244.474512438" observedRunningTime="2025-12-02 10:39:52.285853108 +0000 UTC m=+1245.615991968" watchObservedRunningTime="2025-12-02 10:39:52.292133479 +0000 UTC m=+1245.622272339" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.296815 4679 scope.go:117] "RemoveContainer" containerID="9d2e16681d29edeb6ff57404bb1449758523eb21ad126002b7c7adab173846a8" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.322294 4679 scope.go:117] "RemoveContainer" containerID="14ab5f1bb9777a5f1133e5ff3a96419992da3f4c5705aa8c647232d8561b2857" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.322727 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.332003 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.350853 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.360409 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.365374 4679 scope.go:117] "RemoveContainer" containerID="d6956a4f4f3d5e2f0aac5215de4a6bb8e5102fec5f0ab6cad1d61e9197a598bf" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.373462 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.374087 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="sg-core" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.374178 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="sg-core" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.374281 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="proxy-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.374391 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="proxy-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.374497 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-log" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.374590 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-log" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.374674 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-central-agent" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.374745 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-central-agent" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.374885 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-log" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.374959 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-log" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.375036 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-notification-agent" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.375118 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-notification-agent" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.375195 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.375288 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: E1202 10:39:52.375428 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.375503 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.375839 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-log" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.375947 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="sg-core" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.376029 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" containerName="glance-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.376105 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-log" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.376194 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-notification-agent" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.376289 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="ceilometer-central-agent" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.376406 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" containerName="glance-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.376494 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" containerName="proxy-httpd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.384790 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.387618 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.390430 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.395673 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.404527 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.417697 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.462966 4679 scope.go:117] "RemoveContainer" containerID="8c7f669e05352f9712222cdbe0ad49be5385a980e06c61e53e2a95a934486904" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.484652 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-run-httpd\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.485014 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59l7s\" (UniqueName: \"kubernetes.io/projected/10bd7721-060f-48b4-9931-83960ffe14dd-kube-api-access-59l7s\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.485084 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-config-data\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.485127 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-log-httpd\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.485222 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.485258 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-scripts\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.485339 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.493488 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.495227 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.497780 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.497813 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.497835 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.497851 4679 scope.go:117] "RemoveContainer" containerID="a2c1e359e1d0038891a7043726146cd0875dec70ec23bd403e0a37eec1b35286" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.497923 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-48wtd" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.505164 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.507324 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.515939 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.516561 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.516624 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.521202 4679 scope.go:117] "RemoveContainer" containerID="b36b75676ab9c393bc54e903d245097061243d086034ab24a6f56ca3f1bebd72" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.524218 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.546516 4679 scope.go:117] "RemoveContainer" containerID="b87d66c33269519376e092d1b0bf22a0441cc598188d0a792f9a159942fc33bc" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591015 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591074 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-scripts\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591120 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591178 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-run-httpd\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591235 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59l7s\" (UniqueName: \"kubernetes.io/projected/10bd7721-060f-48b4-9931-83960ffe14dd-kube-api-access-59l7s\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591259 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-config-data\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591287 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-log-httpd\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591756 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-log-httpd\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.591966 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-run-httpd\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.603294 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-config-data\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.604300 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.607519 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.612080 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59l7s\" (UniqueName: \"kubernetes.io/projected/10bd7721-060f-48b4-9931-83960ffe14dd-kube-api-access-59l7s\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.632330 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-scripts\") pod \"ceilometer-0\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693152 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/daa65ea1-139c-49c6-9f88-36efbb79d52c-logs\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693191 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-scripts\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693214 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-config-data\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693247 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693270 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g66g9\" (UniqueName: \"kubernetes.io/projected/4386b4f2-7580-4c18-ae6a-078f8de0edc5-kube-api-access-g66g9\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693295 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693339 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4386b4f2-7580-4c18-ae6a-078f8de0edc5-logs\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693365 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693404 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4386b4f2-7580-4c18-ae6a-078f8de0edc5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693446 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693470 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/daa65ea1-139c-49c6-9f88-36efbb79d52c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693498 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693535 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693559 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693607 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt56t\" (UniqueName: \"kubernetes.io/projected/daa65ea1-139c-49c6-9f88-36efbb79d52c-kube-api-access-gt56t\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.693624 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.704905 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.797922 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4386b4f2-7580-4c18-ae6a-078f8de0edc5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798253 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798283 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/daa65ea1-139c-49c6-9f88-36efbb79d52c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798329 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798384 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798411 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798458 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt56t\" (UniqueName: \"kubernetes.io/projected/daa65ea1-139c-49c6-9f88-36efbb79d52c-kube-api-access-gt56t\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798482 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798496 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4386b4f2-7580-4c18-ae6a-078f8de0edc5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798517 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/daa65ea1-139c-49c6-9f88-36efbb79d52c-logs\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798563 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-scripts\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798606 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-config-data\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798680 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798701 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g66g9\" (UniqueName: \"kubernetes.io/projected/4386b4f2-7580-4c18-ae6a-078f8de0edc5-kube-api-access-g66g9\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798718 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798957 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/daa65ea1-139c-49c6-9f88-36efbb79d52c-logs\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798957 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.798740 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.800328 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4386b4f2-7580-4c18-ae6a-078f8de0edc5-logs\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.800382 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.800715 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4386b4f2-7580-4c18-ae6a-078f8de0edc5-logs\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.800760 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/daa65ea1-139c-49c6-9f88-36efbb79d52c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.805514 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.805924 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-scripts\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.809983 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.827876 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.828253 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt56t\" (UniqueName: \"kubernetes.io/projected/daa65ea1-139c-49c6-9f88-36efbb79d52c-kube-api-access-gt56t\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.831587 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.833764 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.837603 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4386b4f2-7580-4c18-ae6a-078f8de0edc5-config-data\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.838229 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g66g9\" (UniqueName: \"kubernetes.io/projected/4386b4f2-7580-4c18-ae6a-078f8de0edc5-kube-api-access-g66g9\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.838895 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/daa65ea1-139c-49c6-9f88-36efbb79d52c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.865898 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"daa65ea1-139c-49c6-9f88-36efbb79d52c\") " pod="openstack/glance-default-internal-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.872851 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4386b4f2-7580-4c18-ae6a-078f8de0edc5\") " pod="openstack/glance-default-external-api-0" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.928787 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27679d46-4a68-4372-be27-9a9ceb14ac61" path="/var/lib/kubelet/pods/27679d46-4a68-4372-be27-9a9ceb14ac61/volumes" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.930026 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fae901a-a069-4931-adad-b71b0cd4df8e" path="/var/lib/kubelet/pods/8fae901a-a069-4931-adad-b71b0cd4df8e/volumes" Dec 02 10:39:52 crc kubenswrapper[4679]: I1202 10:39:52.931668 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5799263-7d7f-4195-b6b2-70d9315e3a3f" path="/var/lib/kubelet/pods/c5799263-7d7f-4195-b6b2-70d9315e3a3f/volumes" Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.122773 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.130978 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.255034 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.711751 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 10:39:53 crc kubenswrapper[4679]: W1202 10:39:53.720792 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4386b4f2_7580_4c18_ae6a_078f8de0edc5.slice/crio-49951aa6e189e648fa350f15f8903ae0bd1f936bd41c9940a9d7014123fb7d74 WatchSource:0}: Error finding container 49951aa6e189e648fa350f15f8903ae0bd1f936bd41c9940a9d7014123fb7d74: Status 404 returned error can't find the container with id 49951aa6e189e648fa350f15f8903ae0bd1f936bd41c9940a9d7014123fb7d74 Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.807787 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 10:39:53 crc kubenswrapper[4679]: W1202 10:39:53.810399 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaa65ea1_139c_49c6_9f88_36efbb79d52c.slice/crio-5ec989bef5f9734fb998c0b765bd17888c54e257cd50b348647c68794bd0763f WatchSource:0}: Error finding container 5ec989bef5f9734fb998c0b765bd17888c54e257cd50b348647c68794bd0763f: Status 404 returned error can't find the container with id 5ec989bef5f9734fb998c0b765bd17888c54e257cd50b348647c68794bd0763f Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.824954 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:53 crc kubenswrapper[4679]: I1202 10:39:53.825815 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8d8658b7-4bkfp" Dec 02 10:39:54 crc kubenswrapper[4679]: I1202 10:39:54.321320 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"daa65ea1-139c-49c6-9f88-36efbb79d52c","Type":"ContainerStarted","Data":"5ec989bef5f9734fb998c0b765bd17888c54e257cd50b348647c68794bd0763f"} Dec 02 10:39:54 crc kubenswrapper[4679]: I1202 10:39:54.322709 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4386b4f2-7580-4c18-ae6a-078f8de0edc5","Type":"ContainerStarted","Data":"49951aa6e189e648fa350f15f8903ae0bd1f936bd41c9940a9d7014123fb7d74"} Dec 02 10:39:54 crc kubenswrapper[4679]: I1202 10:39:54.326351 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerStarted","Data":"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663"} Dec 02 10:39:54 crc kubenswrapper[4679]: I1202 10:39:54.326385 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerStarted","Data":"1ef6bed5a72d000ba42073d5001aaf5f26a7daa42daa8c607107f4f6ddce566a"} Dec 02 10:39:54 crc kubenswrapper[4679]: I1202 10:39:54.479378 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.297699 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6bd7d57448-84dtd" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.298382 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.336887 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"daa65ea1-139c-49c6-9f88-36efbb79d52c","Type":"ContainerStarted","Data":"87293f12c470a868597e1d9a1ef67d2135802e448afd78f97ab188cfcebd6dd4"} Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.336935 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"daa65ea1-139c-49c6-9f88-36efbb79d52c","Type":"ContainerStarted","Data":"421e004d1586189d0811ab4b828cdc3106910a8e4b80015b86b0cb0bfb828770"} Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.344581 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4386b4f2-7580-4c18-ae6a-078f8de0edc5","Type":"ContainerStarted","Data":"06cabdb7400193e40fb2e49c147eb18ebd25e472986eceba6b43ecd60b516840"} Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.344629 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4386b4f2-7580-4c18-ae6a-078f8de0edc5","Type":"ContainerStarted","Data":"eb64b8487e557615a2d1e5bd1e0719fa57f75a7be99ae59b5108d44c87ceb30e"} Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.346727 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerStarted","Data":"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837"} Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.377959 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.377941809 podStartE2EDuration="3.377941809s" podCreationTimestamp="2025-12-02 10:39:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:55.363545724 +0000 UTC m=+1248.693684584" watchObservedRunningTime="2025-12-02 10:39:55.377941809 +0000 UTC m=+1248.708080669" Dec 02 10:39:55 crc kubenswrapper[4679]: I1202 10:39:55.413521 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.413497995 podStartE2EDuration="3.413497995s" podCreationTimestamp="2025-12-02 10:39:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:39:55.399398718 +0000 UTC m=+1248.729537578" watchObservedRunningTime="2025-12-02 10:39:55.413497995 +0000 UTC m=+1248.743636855" Dec 02 10:39:56 crc kubenswrapper[4679]: I1202 10:39:56.365183 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerStarted","Data":"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66"} Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.386345 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerStarted","Data":"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc"} Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.386749 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.386638 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="sg-core" containerID="cri-o://c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" gracePeriod=30 Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.386484 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="proxy-httpd" containerID="cri-o://5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" gracePeriod=30 Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.386672 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-notification-agent" containerID="cri-o://f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" gracePeriod=30 Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.386699 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-central-agent" containerID="cri-o://7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" gracePeriod=30 Dec 02 10:39:58 crc kubenswrapper[4679]: I1202 10:39:58.419218 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.143348954 podStartE2EDuration="6.419196095s" podCreationTimestamp="2025-12-02 10:39:52 +0000 UTC" firstStartedPulling="2025-12-02 10:39:53.317641203 +0000 UTC m=+1246.647780063" lastFinishedPulling="2025-12-02 10:39:57.593488344 +0000 UTC m=+1250.923627204" observedRunningTime="2025-12-02 10:39:58.409888066 +0000 UTC m=+1251.740026936" watchObservedRunningTime="2025-12-02 10:39:58.419196095 +0000 UTC m=+1251.749334965" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.152609 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.238826 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-run-httpd\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.238896 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-log-httpd\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.238954 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-sg-core-conf-yaml\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.238979 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59l7s\" (UniqueName: \"kubernetes.io/projected/10bd7721-060f-48b4-9931-83960ffe14dd-kube-api-access-59l7s\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.239068 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-config-data\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.239103 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-combined-ca-bundle\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.239232 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-scripts\") pod \"10bd7721-060f-48b4-9931-83960ffe14dd\" (UID: \"10bd7721-060f-48b4-9931-83960ffe14dd\") " Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.239280 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.239670 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.240374 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.244395 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-scripts" (OuterVolumeSpecName: "scripts") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.255066 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10bd7721-060f-48b4-9931-83960ffe14dd-kube-api-access-59l7s" (OuterVolumeSpecName: "kube-api-access-59l7s") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "kube-api-access-59l7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.265575 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.338875 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.341480 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.341512 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/10bd7721-060f-48b4-9931-83960ffe14dd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.341521 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.341532 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59l7s\" (UniqueName: \"kubernetes.io/projected/10bd7721-060f-48b4-9931-83960ffe14dd-kube-api-access-59l7s\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.341541 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.344490 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-config-data" (OuterVolumeSpecName: "config-data") pod "10bd7721-060f-48b4-9931-83960ffe14dd" (UID: "10bd7721-060f-48b4-9931-83960ffe14dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400259 4679 generic.go:334] "Generic (PLEG): container finished" podID="10bd7721-060f-48b4-9931-83960ffe14dd" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" exitCode=0 Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400295 4679 generic.go:334] "Generic (PLEG): container finished" podID="10bd7721-060f-48b4-9931-83960ffe14dd" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" exitCode=2 Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400321 4679 generic.go:334] "Generic (PLEG): container finished" podID="10bd7721-060f-48b4-9931-83960ffe14dd" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" exitCode=0 Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400331 4679 generic.go:334] "Generic (PLEG): container finished" podID="10bd7721-060f-48b4-9931-83960ffe14dd" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" exitCode=0 Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400326 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerDied","Data":"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc"} Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400398 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerDied","Data":"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66"} Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400418 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerDied","Data":"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837"} Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400422 4679 scope.go:117] "RemoveContainer" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400430 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerDied","Data":"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663"} Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400443 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"10bd7721-060f-48b4-9931-83960ffe14dd","Type":"ContainerDied","Data":"1ef6bed5a72d000ba42073d5001aaf5f26a7daa42daa8c607107f4f6ddce566a"} Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.400393 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.429457 4679 scope.go:117] "RemoveContainer" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.440012 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.443458 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10bd7721-060f-48b4-9931-83960ffe14dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.450430 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.453392 4679 scope.go:117] "RemoveContainer" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.468460 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.468943 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-central-agent" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.468967 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-central-agent" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.468983 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-notification-agent" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.468993 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-notification-agent" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.469017 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="proxy-httpd" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.472460 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="proxy-httpd" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.472538 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="sg-core" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.472547 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="sg-core" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.472888 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-central-agent" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.472912 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="sg-core" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.472939 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="ceilometer-notification-agent" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.472948 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" containerName="proxy-httpd" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.474995 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.477016 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.477205 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.484694 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.486299 4679 scope.go:117] "RemoveContainer" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.513552 4679 scope.go:117] "RemoveContainer" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.514228 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": container with ID starting with 5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc not found: ID does not exist" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.514269 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc"} err="failed to get container status \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": rpc error: code = NotFound desc = could not find container \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": container with ID starting with 5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.514295 4679 scope.go:117] "RemoveContainer" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.514727 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": container with ID starting with c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66 not found: ID does not exist" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.514753 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66"} err="failed to get container status \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": rpc error: code = NotFound desc = could not find container \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": container with ID starting with c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.514773 4679 scope.go:117] "RemoveContainer" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.515112 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": container with ID starting with f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837 not found: ID does not exist" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.515141 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837"} err="failed to get container status \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": rpc error: code = NotFound desc = could not find container \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": container with ID starting with f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.515159 4679 scope.go:117] "RemoveContainer" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" Dec 02 10:39:59 crc kubenswrapper[4679]: E1202 10:39:59.515608 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": container with ID starting with 7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663 not found: ID does not exist" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.515644 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663"} err="failed to get container status \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": rpc error: code = NotFound desc = could not find container \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": container with ID starting with 7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.515673 4679 scope.go:117] "RemoveContainer" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.516074 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc"} err="failed to get container status \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": rpc error: code = NotFound desc = could not find container \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": container with ID starting with 5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.516111 4679 scope.go:117] "RemoveContainer" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.516568 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66"} err="failed to get container status \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": rpc error: code = NotFound desc = could not find container \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": container with ID starting with c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.516607 4679 scope.go:117] "RemoveContainer" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.516933 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837"} err="failed to get container status \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": rpc error: code = NotFound desc = could not find container \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": container with ID starting with f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.516963 4679 scope.go:117] "RemoveContainer" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.517372 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663"} err="failed to get container status \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": rpc error: code = NotFound desc = could not find container \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": container with ID starting with 7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.517418 4679 scope.go:117] "RemoveContainer" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.517702 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc"} err="failed to get container status \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": rpc error: code = NotFound desc = could not find container \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": container with ID starting with 5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.517725 4679 scope.go:117] "RemoveContainer" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.519652 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66"} err="failed to get container status \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": rpc error: code = NotFound desc = could not find container \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": container with ID starting with c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.519682 4679 scope.go:117] "RemoveContainer" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.520276 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837"} err="failed to get container status \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": rpc error: code = NotFound desc = could not find container \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": container with ID starting with f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.520327 4679 scope.go:117] "RemoveContainer" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.522451 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663"} err="failed to get container status \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": rpc error: code = NotFound desc = could not find container \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": container with ID starting with 7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.522492 4679 scope.go:117] "RemoveContainer" containerID="5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.525547 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc"} err="failed to get container status \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": rpc error: code = NotFound desc = could not find container \"5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc\": container with ID starting with 5459b378b1fe56e67fbdc81083eb31b09943b307b56384cfae213168c9247ffc not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.525585 4679 scope.go:117] "RemoveContainer" containerID="c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.525999 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66"} err="failed to get container status \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": rpc error: code = NotFound desc = could not find container \"c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66\": container with ID starting with c19f6a7827b7b86a01d4366adf58e868fbc8319a08a0d60bfe3f13cbd9916b66 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.526025 4679 scope.go:117] "RemoveContainer" containerID="f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.526219 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837"} err="failed to get container status \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": rpc error: code = NotFound desc = could not find container \"f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837\": container with ID starting with f37d2a50bbfed33853ed47ea7c70ee7eebb48e51a06faba7927aca25c31aa837 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.526240 4679 scope.go:117] "RemoveContainer" containerID="7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.526470 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663"} err="failed to get container status \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": rpc error: code = NotFound desc = could not find container \"7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663\": container with ID starting with 7d63913649508f22aaace525fc7b5205c476ce0cbac355d02fd411ef1b2b8663 not found: ID does not exist" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646618 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-config-data\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646672 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646725 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-log-httpd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646782 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pntbd\" (UniqueName: \"kubernetes.io/projected/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-kube-api-access-pntbd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646843 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-run-httpd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646862 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.646891 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-scripts\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.748944 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.748997 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-scripts\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749087 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-config-data\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749112 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749163 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-log-httpd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749191 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pntbd\" (UniqueName: \"kubernetes.io/projected/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-kube-api-access-pntbd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749255 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-run-httpd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749797 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-run-httpd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.749833 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-log-httpd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.753361 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.753415 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-scripts\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.754861 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-config-data\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.755363 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.770722 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pntbd\" (UniqueName: \"kubernetes.io/projected/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-kube-api-access-pntbd\") pod \"ceilometer-0\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.801457 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:39:59 crc kubenswrapper[4679]: I1202 10:39:59.952085 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.289651 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:00 crc kubenswrapper[4679]: W1202 10:40:00.308967 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fc553c8_18cf_4d05_a449_7ac987c8eb5d.slice/crio-1ec9880d2a5d88709431e5bfa2fdd5ae8332a4abbc4b3f0c627f886665406d3d WatchSource:0}: Error finding container 1ec9880d2a5d88709431e5bfa2fdd5ae8332a4abbc4b3f0c627f886665406d3d: Status 404 returned error can't find the container with id 1ec9880d2a5d88709431e5bfa2fdd5ae8332a4abbc4b3f0c627f886665406d3d Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.413802 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerStarted","Data":"1ec9880d2a5d88709431e5bfa2fdd5ae8332a4abbc4b3f0c627f886665406d3d"} Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.415871 4679 generic.go:334] "Generic (PLEG): container finished" podID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerID="d87f86cc90eb58eaa5a8aa14c2723787b723ae1ed4cfb00f1c88c0c427250f59" exitCode=137 Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.415925 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bd7d57448-84dtd" event={"ID":"ded5eaca-d45b-46bf-a770-50d909f0003e","Type":"ContainerDied","Data":"d87f86cc90eb58eaa5a8aa14c2723787b723ae1ed4cfb00f1c88c0c427250f59"} Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.758156 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.870827 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-config-data\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871183 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-combined-ca-bundle\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871245 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-secret-key\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871290 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-tls-certs\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871355 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn297\" (UniqueName: \"kubernetes.io/projected/ded5eaca-d45b-46bf-a770-50d909f0003e-kube-api-access-bn297\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871440 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded5eaca-d45b-46bf-a770-50d909f0003e-logs\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871554 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-scripts\") pod \"ded5eaca-d45b-46bf-a770-50d909f0003e\" (UID: \"ded5eaca-d45b-46bf-a770-50d909f0003e\") " Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.871907 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ded5eaca-d45b-46bf-a770-50d909f0003e-logs" (OuterVolumeSpecName: "logs") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.877457 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.877783 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ded5eaca-d45b-46bf-a770-50d909f0003e-kube-api-access-bn297" (OuterVolumeSpecName: "kube-api-access-bn297") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "kube-api-access-bn297". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.900826 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-scripts" (OuterVolumeSpecName: "scripts") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.902116 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.904736 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-config-data" (OuterVolumeSpecName: "config-data") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.919695 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10bd7721-060f-48b4-9931-83960ffe14dd" path="/var/lib/kubelet/pods/10bd7721-060f-48b4-9931-83960ffe14dd/volumes" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.927810 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "ded5eaca-d45b-46bf-a770-50d909f0003e" (UID: "ded5eaca-d45b-46bf-a770-50d909f0003e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.973673 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.973807 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ded5eaca-d45b-46bf-a770-50d909f0003e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.973864 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.973938 4679 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.973999 4679 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ded5eaca-d45b-46bf-a770-50d909f0003e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.974052 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn297\" (UniqueName: \"kubernetes.io/projected/ded5eaca-d45b-46bf-a770-50d909f0003e-kube-api-access-bn297\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:00 crc kubenswrapper[4679]: I1202 10:40:00.974102 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ded5eaca-d45b-46bf-a770-50d909f0003e-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.427408 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerStarted","Data":"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51"} Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.429238 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bd7d57448-84dtd" event={"ID":"ded5eaca-d45b-46bf-a770-50d909f0003e","Type":"ContainerDied","Data":"9ddb76da0ced0815c162346f9bf7c3e84a84e82dc2aa9d29019c4c5218bedada"} Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.429285 4679 scope.go:117] "RemoveContainer" containerID="dbbabf0a29747ebfd44364b9b3f7b60a84d413aee3ffe565fee42f380f0ae98f" Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.429428 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bd7d57448-84dtd" Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.461233 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bd7d57448-84dtd"] Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.470640 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6bd7d57448-84dtd"] Dec 02 10:40:01 crc kubenswrapper[4679]: I1202 10:40:01.606333 4679 scope.go:117] "RemoveContainer" containerID="d87f86cc90eb58eaa5a8aa14c2723787b723ae1ed4cfb00f1c88c0c427250f59" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.439975 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerStarted","Data":"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf"} Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.920587 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" path="/var/lib/kubelet/pods/ded5eaca-d45b-46bf-a770-50d909f0003e/volumes" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.921271 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lvb5d"] Dec 02 10:40:02 crc kubenswrapper[4679]: E1202 10:40:02.921602 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon-log" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.921618 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon-log" Dec 02 10:40:02 crc kubenswrapper[4679]: E1202 10:40:02.921656 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.921664 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.921834 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.921870 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded5eaca-d45b-46bf-a770-50d909f0003e" containerName="horizon-log" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.922528 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:02 crc kubenswrapper[4679]: I1202 10:40:02.938587 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lvb5d"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.015973 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st9vz\" (UniqueName: \"kubernetes.io/projected/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-kube-api-access-st9vz\") pod \"nova-api-db-create-lvb5d\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.016335 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-operator-scripts\") pod \"nova-api-db-create-lvb5d\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.026169 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-fj9wm"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.027638 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.037888 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-dbb7-account-create-update-mc5lj"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.039701 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.042169 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.057984 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-fj9wm"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.074195 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-dbb7-account-create-update-mc5lj"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.117696 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrlgj\" (UniqueName: \"kubernetes.io/projected/5bc64597-c029-4073-8f2f-9d25583c1767-kube-api-access-wrlgj\") pod \"nova-cell0-db-create-fj9wm\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.118241 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30145675-4813-4416-a06e-1487bf15c0fe-operator-scripts\") pod \"nova-api-dbb7-account-create-update-mc5lj\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.118357 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st9vz\" (UniqueName: \"kubernetes.io/projected/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-kube-api-access-st9vz\") pod \"nova-api-db-create-lvb5d\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.118392 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpjrw\" (UniqueName: \"kubernetes.io/projected/30145675-4813-4416-a06e-1487bf15c0fe-kube-api-access-hpjrw\") pod \"nova-api-dbb7-account-create-update-mc5lj\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.118431 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-operator-scripts\") pod \"nova-api-db-create-lvb5d\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.118460 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bc64597-c029-4073-8f2f-9d25583c1767-operator-scripts\") pod \"nova-cell0-db-create-fj9wm\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.119358 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-operator-scripts\") pod \"nova-api-db-create-lvb5d\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.124745 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.124782 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.131281 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.131317 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.142981 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st9vz\" (UniqueName: \"kubernetes.io/projected/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-kube-api-access-st9vz\") pod \"nova-api-db-create-lvb5d\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.178150 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.179509 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.179818 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.186922 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.219796 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bc64597-c029-4073-8f2f-9d25583c1767-operator-scripts\") pod \"nova-cell0-db-create-fj9wm\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.219905 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrlgj\" (UniqueName: \"kubernetes.io/projected/5bc64597-c029-4073-8f2f-9d25583c1767-kube-api-access-wrlgj\") pod \"nova-cell0-db-create-fj9wm\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.220022 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30145675-4813-4416-a06e-1487bf15c0fe-operator-scripts\") pod \"nova-api-dbb7-account-create-update-mc5lj\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.220100 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpjrw\" (UniqueName: \"kubernetes.io/projected/30145675-4813-4416-a06e-1487bf15c0fe-kube-api-access-hpjrw\") pod \"nova-api-dbb7-account-create-update-mc5lj\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.221746 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bc64597-c029-4073-8f2f-9d25583c1767-operator-scripts\") pod \"nova-cell0-db-create-fj9wm\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.221746 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30145675-4813-4416-a06e-1487bf15c0fe-operator-scripts\") pod \"nova-api-dbb7-account-create-update-mc5lj\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.240343 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-feeb-account-create-update-v5pzb"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.241567 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.241705 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.250777 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.252509 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrlgj\" (UniqueName: \"kubernetes.io/projected/5bc64597-c029-4073-8f2f-9d25583c1767-kube-api-access-wrlgj\") pod \"nova-cell0-db-create-fj9wm\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.253490 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpjrw\" (UniqueName: \"kubernetes.io/projected/30145675-4813-4416-a06e-1487bf15c0fe-kube-api-access-hpjrw\") pod \"nova-api-dbb7-account-create-update-mc5lj\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.260235 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b5tpv"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.261521 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.270965 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b5tpv"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.282396 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-feeb-account-create-update-v5pzb"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.322880 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kwz2\" (UniqueName: \"kubernetes.io/projected/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-kube-api-access-2kwz2\") pod \"nova-cell0-feeb-account-create-update-v5pzb\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.322988 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccfwq\" (UniqueName: \"kubernetes.io/projected/0d1a8e7c-e69d-4400-84c2-27537b0593f0-kube-api-access-ccfwq\") pod \"nova-cell1-db-create-b5tpv\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.323022 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-operator-scripts\") pod \"nova-cell0-feeb-account-create-update-v5pzb\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.323159 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a8e7c-e69d-4400-84c2-27537b0593f0-operator-scripts\") pod \"nova-cell1-db-create-b5tpv\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.348964 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.361814 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.427074 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kwz2\" (UniqueName: \"kubernetes.io/projected/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-kube-api-access-2kwz2\") pod \"nova-cell0-feeb-account-create-update-v5pzb\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.427142 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccfwq\" (UniqueName: \"kubernetes.io/projected/0d1a8e7c-e69d-4400-84c2-27537b0593f0-kube-api-access-ccfwq\") pod \"nova-cell1-db-create-b5tpv\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.427161 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-operator-scripts\") pod \"nova-cell0-feeb-account-create-update-v5pzb\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.427220 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a8e7c-e69d-4400-84c2-27537b0593f0-operator-scripts\") pod \"nova-cell1-db-create-b5tpv\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.428015 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a8e7c-e69d-4400-84c2-27537b0593f0-operator-scripts\") pod \"nova-cell1-db-create-b5tpv\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.434224 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-operator-scripts\") pod \"nova-cell0-feeb-account-create-update-v5pzb\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.435754 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4f48-account-create-update-5kwkr"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.437558 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.440458 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.453200 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccfwq\" (UniqueName: \"kubernetes.io/projected/0d1a8e7c-e69d-4400-84c2-27537b0593f0-kube-api-access-ccfwq\") pod \"nova-cell1-db-create-b5tpv\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.453476 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4f48-account-create-update-5kwkr"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.474163 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kwz2\" (UniqueName: \"kubernetes.io/projected/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-kube-api-access-2kwz2\") pod \"nova-cell0-feeb-account-create-update-v5pzb\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.510244 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.523588 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.523712 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.523725 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.523737 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.541163 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl979\" (UniqueName: \"kubernetes.io/projected/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-kube-api-access-fl979\") pod \"nova-cell1-4f48-account-create-update-5kwkr\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.541460 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-operator-scripts\") pod \"nova-cell1-4f48-account-create-update-5kwkr\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.643703 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-operator-scripts\") pod \"nova-cell1-4f48-account-create-update-5kwkr\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.644030 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl979\" (UniqueName: \"kubernetes.io/projected/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-kube-api-access-fl979\") pod \"nova-cell1-4f48-account-create-update-5kwkr\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.646770 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-operator-scripts\") pod \"nova-cell1-4f48-account-create-update-5kwkr\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.671087 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl979\" (UniqueName: \"kubernetes.io/projected/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-kube-api-access-fl979\") pod \"nova-cell1-4f48-account-create-update-5kwkr\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.672260 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.788991 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lvb5d"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.812599 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.919363 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-dbb7-account-create-update-mc5lj"] Dec 02 10:40:03 crc kubenswrapper[4679]: I1202 10:40:03.945999 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-fj9wm"] Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.119776 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b5tpv"] Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.218197 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-feeb-account-create-update-v5pzb"] Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.348317 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4f48-account-create-update-5kwkr"] Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.530492 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" event={"ID":"afead6b9-e07e-4d06-af9a-2f876bb4c5e8","Type":"ContainerStarted","Data":"5607a9f1de796bac6f2af3341c6a169797fec5bb617bdad598a2bdf5ba6eb42f"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.532239 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" event={"ID":"c372bf97-ea32-4ca7-99cc-9ead0f8e277f","Type":"ContainerStarted","Data":"c6c43ba1e7c3a92d09f7a1b2524a699d66fab11ce00e03d08a57bda981a4e690"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.533771 4679 generic.go:334] "Generic (PLEG): container finished" podID="30145675-4813-4416-a06e-1487bf15c0fe" containerID="4731d335898e37ad10b75005bac9e8b7df4021a8932215936289081b270620ee" exitCode=0 Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.533817 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" event={"ID":"30145675-4813-4416-a06e-1487bf15c0fe","Type":"ContainerDied","Data":"4731d335898e37ad10b75005bac9e8b7df4021a8932215936289081b270620ee"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.533835 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" event={"ID":"30145675-4813-4416-a06e-1487bf15c0fe","Type":"ContainerStarted","Data":"f50b4aeb5a1aee25dd0ad1254b80e2df0ab6a9c3b5d2788086c6f424fa1c649f"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.535481 4679 generic.go:334] "Generic (PLEG): container finished" podID="89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" containerID="aed740d8d2ec18baeaea3f61355f29c1cbcedf2576bafb2b36c24b5aef0f4c07" exitCode=0 Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.535527 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lvb5d" event={"ID":"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3","Type":"ContainerDied","Data":"aed740d8d2ec18baeaea3f61355f29c1cbcedf2576bafb2b36c24b5aef0f4c07"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.535541 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lvb5d" event={"ID":"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3","Type":"ContainerStarted","Data":"79414f4e7f92f7e7d2696766a125da475d6ec2fdd3a40d393020859cc8686579"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.537050 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5tpv" event={"ID":"0d1a8e7c-e69d-4400-84c2-27537b0593f0","Type":"ContainerStarted","Data":"11b6e716eba359b44071efaba7bae07b1230ca79aa0fee91e8a147d855444a02"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.545821 4679 generic.go:334] "Generic (PLEG): container finished" podID="5bc64597-c029-4073-8f2f-9d25583c1767" containerID="d3bd56934222bcfb7262c2adcfaa69189e6ad959c857cf212711ec2ee2a820ff" exitCode=0 Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.545888 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fj9wm" event={"ID":"5bc64597-c029-4073-8f2f-9d25583c1767","Type":"ContainerDied","Data":"d3bd56934222bcfb7262c2adcfaa69189e6ad959c857cf212711ec2ee2a820ff"} Dec 02 10:40:04 crc kubenswrapper[4679]: I1202 10:40:04.545920 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fj9wm" event={"ID":"5bc64597-c029-4073-8f2f-9d25583c1767","Type":"ContainerStarted","Data":"f017b896a133b27a4dbe841e0ad46af71cf48d71c9c98077bd401441d94b56f0"} Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.560991 4679 generic.go:334] "Generic (PLEG): container finished" podID="0d1a8e7c-e69d-4400-84c2-27537b0593f0" containerID="c9dc10bfade2c586c041c88061a89a592c6213c7d802b2e701d808de41573cd9" exitCode=0 Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.561076 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5tpv" event={"ID":"0d1a8e7c-e69d-4400-84c2-27537b0593f0","Type":"ContainerDied","Data":"c9dc10bfade2c586c041c88061a89a592c6213c7d802b2e701d808de41573cd9"} Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.563513 4679 generic.go:334] "Generic (PLEG): container finished" podID="afead6b9-e07e-4d06-af9a-2f876bb4c5e8" containerID="2348b19a4bc333500eeda889787e776da3d58468521f673fa9a305dd77b96998" exitCode=0 Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.563571 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" event={"ID":"afead6b9-e07e-4d06-af9a-2f876bb4c5e8","Type":"ContainerDied","Data":"2348b19a4bc333500eeda889787e776da3d58468521f673fa9a305dd77b96998"} Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.566889 4679 generic.go:334] "Generic (PLEG): container finished" podID="c372bf97-ea32-4ca7-99cc-9ead0f8e277f" containerID="a8014e75c6e0af69dda28c0fe97bc90313f7c37f211b5af08a20400efabcb901" exitCode=0 Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.566918 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" event={"ID":"c372bf97-ea32-4ca7-99cc-9ead0f8e277f","Type":"ContainerDied","Data":"a8014e75c6e0af69dda28c0fe97bc90313f7c37f211b5af08a20400efabcb901"} Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.751977 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.752090 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.758605 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.812213 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.812332 4679 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 10:40:05 crc kubenswrapper[4679]: I1202 10:40:05.812988 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.110676 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.271767 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.278773 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.304511 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30145675-4813-4416-a06e-1487bf15c0fe-operator-scripts\") pod \"30145675-4813-4416-a06e-1487bf15c0fe\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.305007 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpjrw\" (UniqueName: \"kubernetes.io/projected/30145675-4813-4416-a06e-1487bf15c0fe-kube-api-access-hpjrw\") pod \"30145675-4813-4416-a06e-1487bf15c0fe\" (UID: \"30145675-4813-4416-a06e-1487bf15c0fe\") " Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.305397 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30145675-4813-4416-a06e-1487bf15c0fe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30145675-4813-4416-a06e-1487bf15c0fe" (UID: "30145675-4813-4416-a06e-1487bf15c0fe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.305590 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30145675-4813-4416-a06e-1487bf15c0fe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.331582 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30145675-4813-4416-a06e-1487bf15c0fe-kube-api-access-hpjrw" (OuterVolumeSpecName: "kube-api-access-hpjrw") pod "30145675-4813-4416-a06e-1487bf15c0fe" (UID: "30145675-4813-4416-a06e-1487bf15c0fe"). InnerVolumeSpecName "kube-api-access-hpjrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.406858 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st9vz\" (UniqueName: \"kubernetes.io/projected/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-kube-api-access-st9vz\") pod \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407009 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bc64597-c029-4073-8f2f-9d25583c1767-operator-scripts\") pod \"5bc64597-c029-4073-8f2f-9d25583c1767\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407043 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrlgj\" (UniqueName: \"kubernetes.io/projected/5bc64597-c029-4073-8f2f-9d25583c1767-kube-api-access-wrlgj\") pod \"5bc64597-c029-4073-8f2f-9d25583c1767\" (UID: \"5bc64597-c029-4073-8f2f-9d25583c1767\") " Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407082 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-operator-scripts\") pod \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\" (UID: \"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3\") " Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407472 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc64597-c029-4073-8f2f-9d25583c1767-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5bc64597-c029-4073-8f2f-9d25583c1767" (UID: "5bc64597-c029-4073-8f2f-9d25583c1767"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407626 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpjrw\" (UniqueName: \"kubernetes.io/projected/30145675-4813-4416-a06e-1487bf15c0fe-kube-api-access-hpjrw\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407644 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5bc64597-c029-4073-8f2f-9d25583c1767-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.407913 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" (UID: "89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.413527 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc64597-c029-4073-8f2f-9d25583c1767-kube-api-access-wrlgj" (OuterVolumeSpecName: "kube-api-access-wrlgj") pod "5bc64597-c029-4073-8f2f-9d25583c1767" (UID: "5bc64597-c029-4073-8f2f-9d25583c1767"). InnerVolumeSpecName "kube-api-access-wrlgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.413624 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-kube-api-access-st9vz" (OuterVolumeSpecName: "kube-api-access-st9vz") pod "89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" (UID: "89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3"). InnerVolumeSpecName "kube-api-access-st9vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.509342 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrlgj\" (UniqueName: \"kubernetes.io/projected/5bc64597-c029-4073-8f2f-9d25583c1767-kube-api-access-wrlgj\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.509376 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.509385 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st9vz\" (UniqueName: \"kubernetes.io/projected/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3-kube-api-access-st9vz\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.576537 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.576660 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dbb7-account-create-update-mc5lj" event={"ID":"30145675-4813-4416-a06e-1487bf15c0fe","Type":"ContainerDied","Data":"f50b4aeb5a1aee25dd0ad1254b80e2df0ab6a9c3b5d2788086c6f424fa1c649f"} Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.576992 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f50b4aeb5a1aee25dd0ad1254b80e2df0ab6a9c3b5d2788086c6f424fa1c649f" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.579219 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lvb5d" event={"ID":"89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3","Type":"ContainerDied","Data":"79414f4e7f92f7e7d2696766a125da475d6ec2fdd3a40d393020859cc8686579"} Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.579251 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79414f4e7f92f7e7d2696766a125da475d6ec2fdd3a40d393020859cc8686579" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.579298 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lvb5d" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.587957 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fj9wm" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.588664 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fj9wm" event={"ID":"5bc64597-c029-4073-8f2f-9d25583c1767","Type":"ContainerDied","Data":"f017b896a133b27a4dbe841e0ad46af71cf48d71c9c98077bd401441d94b56f0"} Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.588700 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f017b896a133b27a4dbe841e0ad46af71cf48d71c9c98077bd401441d94b56f0" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.975462 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:06 crc kubenswrapper[4679]: I1202 10:40:06.987871 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.030455 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.126649 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-operator-scripts\") pod \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.126707 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kwz2\" (UniqueName: \"kubernetes.io/projected/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-kube-api-access-2kwz2\") pod \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\" (UID: \"c372bf97-ea32-4ca7-99cc-9ead0f8e277f\") " Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.126749 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl979\" (UniqueName: \"kubernetes.io/projected/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-kube-api-access-fl979\") pod \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.126794 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-operator-scripts\") pod \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\" (UID: \"afead6b9-e07e-4d06-af9a-2f876bb4c5e8\") " Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.127625 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afead6b9-e07e-4d06-af9a-2f876bb4c5e8" (UID: "afead6b9-e07e-4d06-af9a-2f876bb4c5e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.127617 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c372bf97-ea32-4ca7-99cc-9ead0f8e277f" (UID: "c372bf97-ea32-4ca7-99cc-9ead0f8e277f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.127670 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccfwq\" (UniqueName: \"kubernetes.io/projected/0d1a8e7c-e69d-4400-84c2-27537b0593f0-kube-api-access-ccfwq\") pod \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.127803 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a8e7c-e69d-4400-84c2-27537b0593f0-operator-scripts\") pod \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\" (UID: \"0d1a8e7c-e69d-4400-84c2-27537b0593f0\") " Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.128402 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.128422 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.128801 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d1a8e7c-e69d-4400-84c2-27537b0593f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d1a8e7c-e69d-4400-84c2-27537b0593f0" (UID: "0d1a8e7c-e69d-4400-84c2-27537b0593f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.132666 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-kube-api-access-2kwz2" (OuterVolumeSpecName: "kube-api-access-2kwz2") pod "c372bf97-ea32-4ca7-99cc-9ead0f8e277f" (UID: "c372bf97-ea32-4ca7-99cc-9ead0f8e277f"). InnerVolumeSpecName "kube-api-access-2kwz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.133120 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d1a8e7c-e69d-4400-84c2-27537b0593f0-kube-api-access-ccfwq" (OuterVolumeSpecName: "kube-api-access-ccfwq") pod "0d1a8e7c-e69d-4400-84c2-27537b0593f0" (UID: "0d1a8e7c-e69d-4400-84c2-27537b0593f0"). InnerVolumeSpecName "kube-api-access-ccfwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.135293 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-kube-api-access-fl979" (OuterVolumeSpecName: "kube-api-access-fl979") pod "afead6b9-e07e-4d06-af9a-2f876bb4c5e8" (UID: "afead6b9-e07e-4d06-af9a-2f876bb4c5e8"). InnerVolumeSpecName "kube-api-access-fl979". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.230617 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccfwq\" (UniqueName: \"kubernetes.io/projected/0d1a8e7c-e69d-4400-84c2-27537b0593f0-kube-api-access-ccfwq\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.230658 4679 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a8e7c-e69d-4400-84c2-27537b0593f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.230674 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kwz2\" (UniqueName: \"kubernetes.io/projected/c372bf97-ea32-4ca7-99cc-9ead0f8e277f-kube-api-access-2kwz2\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.230687 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl979\" (UniqueName: \"kubernetes.io/projected/afead6b9-e07e-4d06-af9a-2f876bb4c5e8-kube-api-access-fl979\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.627061 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.627168 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4f48-account-create-update-5kwkr" event={"ID":"afead6b9-e07e-4d06-af9a-2f876bb4c5e8","Type":"ContainerDied","Data":"5607a9f1de796bac6f2af3341c6a169797fec5bb617bdad598a2bdf5ba6eb42f"} Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.629070 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5607a9f1de796bac6f2af3341c6a169797fec5bb617bdad598a2bdf5ba6eb42f" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.629182 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" event={"ID":"c372bf97-ea32-4ca7-99cc-9ead0f8e277f","Type":"ContainerDied","Data":"c6c43ba1e7c3a92d09f7a1b2524a699d66fab11ce00e03d08a57bda981a4e690"} Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.629197 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6c43ba1e7c3a92d09f7a1b2524a699d66fab11ce00e03d08a57bda981a4e690" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.629253 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-feeb-account-create-update-v5pzb" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.631164 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5tpv" event={"ID":"0d1a8e7c-e69d-4400-84c2-27537b0593f0","Type":"ContainerDied","Data":"11b6e716eba359b44071efaba7bae07b1230ca79aa0fee91e8a147d855444a02"} Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.631298 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11b6e716eba359b44071efaba7bae07b1230ca79aa0fee91e8a147d855444a02" Dec 02 10:40:07 crc kubenswrapper[4679]: I1202 10:40:07.631485 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5tpv" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483073 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bm8fd"] Dec 02 10:40:08 crc kubenswrapper[4679]: E1202 10:40:08.483469 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30145675-4813-4416-a06e-1487bf15c0fe" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483488 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="30145675-4813-4416-a06e-1487bf15c0fe" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: E1202 10:40:08.483498 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c372bf97-ea32-4ca7-99cc-9ead0f8e277f" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483505 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="c372bf97-ea32-4ca7-99cc-9ead0f8e277f" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: E1202 10:40:08.483523 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afead6b9-e07e-4d06-af9a-2f876bb4c5e8" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483531 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="afead6b9-e07e-4d06-af9a-2f876bb4c5e8" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: E1202 10:40:08.483556 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc64597-c029-4073-8f2f-9d25583c1767" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483564 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc64597-c029-4073-8f2f-9d25583c1767" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: E1202 10:40:08.483579 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483586 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: E1202 10:40:08.483601 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d1a8e7c-e69d-4400-84c2-27537b0593f0" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483610 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d1a8e7c-e69d-4400-84c2-27537b0593f0" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483818 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="30145675-4813-4416-a06e-1487bf15c0fe" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483833 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc64597-c029-4073-8f2f-9d25583c1767" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483848 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="c372bf97-ea32-4ca7-99cc-9ead0f8e277f" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483861 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="afead6b9-e07e-4d06-af9a-2f876bb4c5e8" containerName="mariadb-account-create-update" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483875 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d1a8e7c-e69d-4400-84c2-27537b0593f0" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.483898 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" containerName="mariadb-database-create" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.484597 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.486427 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.486686 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.487470 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jr68j" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.495762 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bm8fd"] Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.650787 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerStarted","Data":"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799"} Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.671112 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpkhx\" (UniqueName: \"kubernetes.io/projected/b287fe6f-16df-4020-85b8-21ebdc224103-kube-api-access-qpkhx\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.671550 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.671695 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-scripts\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.671796 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-config-data\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.773015 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpkhx\" (UniqueName: \"kubernetes.io/projected/b287fe6f-16df-4020-85b8-21ebdc224103-kube-api-access-qpkhx\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.773340 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.773444 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-scripts\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.773521 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-config-data\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.778907 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-scripts\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.778999 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.779402 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-config-data\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.790163 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpkhx\" (UniqueName: \"kubernetes.io/projected/b287fe6f-16df-4020-85b8-21ebdc224103-kube-api-access-qpkhx\") pod \"nova-cell0-conductor-db-sync-bm8fd\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:08 crc kubenswrapper[4679]: I1202 10:40:08.802477 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:09 crc kubenswrapper[4679]: W1202 10:40:09.261117 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb287fe6f_16df_4020_85b8_21ebdc224103.slice/crio-dd3282054b10d3a916d64b92c0f67a0717bbd653c8b6fb3b4fb16b3cd717847a WatchSource:0}: Error finding container dd3282054b10d3a916d64b92c0f67a0717bbd653c8b6fb3b4fb16b3cd717847a: Status 404 returned error can't find the container with id dd3282054b10d3a916d64b92c0f67a0717bbd653c8b6fb3b4fb16b3cd717847a Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.261482 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bm8fd"] Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.671581 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" event={"ID":"b287fe6f-16df-4020-85b8-21ebdc224103","Type":"ContainerStarted","Data":"dd3282054b10d3a916d64b92c0f67a0717bbd653c8b6fb3b4fb16b3cd717847a"} Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.674785 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerStarted","Data":"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810"} Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.674899 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-central-agent" containerID="cri-o://6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" gracePeriod=30 Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.674934 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.674980 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-notification-agent" containerID="cri-o://5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" gracePeriod=30 Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.674973 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="sg-core" containerID="cri-o://8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" gracePeriod=30 Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.675089 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="proxy-httpd" containerID="cri-o://40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" gracePeriod=30 Dec 02 10:40:09 crc kubenswrapper[4679]: I1202 10:40:09.700943 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.047165618 podStartE2EDuration="10.700926314s" podCreationTimestamp="2025-12-02 10:39:59 +0000 UTC" firstStartedPulling="2025-12-02 10:40:00.311351181 +0000 UTC m=+1253.641490041" lastFinishedPulling="2025-12-02 10:40:08.965111877 +0000 UTC m=+1262.295250737" observedRunningTime="2025-12-02 10:40:09.69419887 +0000 UTC m=+1263.024337730" watchObservedRunningTime="2025-12-02 10:40:09.700926314 +0000 UTC m=+1263.031065174" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.489326 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512506 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-config-data\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512612 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-scripts\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512717 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-sg-core-conf-yaml\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512783 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pntbd\" (UniqueName: \"kubernetes.io/projected/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-kube-api-access-pntbd\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512844 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-log-httpd\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512888 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-combined-ca-bundle\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.512908 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-run-httpd\") pod \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\" (UID: \"7fc553c8-18cf-4d05-a449-7ac987c8eb5d\") " Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.513727 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.513947 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.514330 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.514373 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.518795 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-scripts" (OuterVolumeSpecName: "scripts") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.520463 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-kube-api-access-pntbd" (OuterVolumeSpecName: "kube-api-access-pntbd") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "kube-api-access-pntbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.553769 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.594576 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.607210 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-config-data" (OuterVolumeSpecName: "config-data") pod "7fc553c8-18cf-4d05-a449-7ac987c8eb5d" (UID: "7fc553c8-18cf-4d05-a449-7ac987c8eb5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.615866 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.615920 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.615932 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.615946 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pntbd\" (UniqueName: \"kubernetes.io/projected/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-kube-api-access-pntbd\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.615958 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc553c8-18cf-4d05-a449-7ac987c8eb5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688604 4679 generic.go:334] "Generic (PLEG): container finished" podID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" exitCode=0 Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688635 4679 generic.go:334] "Generic (PLEG): container finished" podID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" exitCode=2 Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688643 4679 generic.go:334] "Generic (PLEG): container finished" podID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" exitCode=0 Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688651 4679 generic.go:334] "Generic (PLEG): container finished" podID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" exitCode=0 Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688668 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerDied","Data":"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810"} Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688693 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerDied","Data":"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799"} Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688704 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerDied","Data":"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf"} Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688714 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerDied","Data":"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51"} Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688722 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc553c8-18cf-4d05-a449-7ac987c8eb5d","Type":"ContainerDied","Data":"1ec9880d2a5d88709431e5bfa2fdd5ae8332a4abbc4b3f0c627f886665406d3d"} Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688736 4679 scope.go:117] "RemoveContainer" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.688834 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.720638 4679 scope.go:117] "RemoveContainer" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.758389 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.767604 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.777290 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.777834 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-notification-agent" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.777934 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-notification-agent" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.778018 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="proxy-httpd" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.778084 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="proxy-httpd" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.783480 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-central-agent" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.783507 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-central-agent" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.783584 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="sg-core" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.783595 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="sg-core" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.784106 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="sg-core" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.784127 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-notification-agent" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.784230 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="ceilometer-central-agent" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.784252 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" containerName="proxy-httpd" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.786293 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.793223 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.805027 4679 scope.go:117] "RemoveContainer" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.805506 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.805690 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.860574 4679 scope.go:117] "RemoveContainer" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.880517 4679 scope.go:117] "RemoveContainer" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.881089 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": container with ID starting with 40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810 not found: ID does not exist" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.881146 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810"} err="failed to get container status \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": rpc error: code = NotFound desc = could not find container \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": container with ID starting with 40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.881177 4679 scope.go:117] "RemoveContainer" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.881612 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": container with ID starting with 8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799 not found: ID does not exist" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.881654 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799"} err="failed to get container status \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": rpc error: code = NotFound desc = could not find container \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": container with ID starting with 8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.881680 4679 scope.go:117] "RemoveContainer" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.881985 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": container with ID starting with 5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf not found: ID does not exist" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.882006 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf"} err="failed to get container status \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": rpc error: code = NotFound desc = could not find container \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": container with ID starting with 5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.882038 4679 scope.go:117] "RemoveContainer" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" Dec 02 10:40:10 crc kubenswrapper[4679]: E1202 10:40:10.882634 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": container with ID starting with 6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51 not found: ID does not exist" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.882685 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51"} err="failed to get container status \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": rpc error: code = NotFound desc = could not find container \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": container with ID starting with 6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.882723 4679 scope.go:117] "RemoveContainer" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.883178 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810"} err="failed to get container status \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": rpc error: code = NotFound desc = could not find container \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": container with ID starting with 40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.883204 4679 scope.go:117] "RemoveContainer" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.883468 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799"} err="failed to get container status \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": rpc error: code = NotFound desc = could not find container \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": container with ID starting with 8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.883499 4679 scope.go:117] "RemoveContainer" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.883806 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf"} err="failed to get container status \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": rpc error: code = NotFound desc = could not find container \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": container with ID starting with 5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.883834 4679 scope.go:117] "RemoveContainer" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.884286 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51"} err="failed to get container status \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": rpc error: code = NotFound desc = could not find container \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": container with ID starting with 6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.884351 4679 scope.go:117] "RemoveContainer" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.884691 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810"} err="failed to get container status \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": rpc error: code = NotFound desc = could not find container \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": container with ID starting with 40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.884721 4679 scope.go:117] "RemoveContainer" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.885007 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799"} err="failed to get container status \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": rpc error: code = NotFound desc = could not find container \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": container with ID starting with 8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.885034 4679 scope.go:117] "RemoveContainer" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.885340 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf"} err="failed to get container status \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": rpc error: code = NotFound desc = could not find container \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": container with ID starting with 5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.885369 4679 scope.go:117] "RemoveContainer" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.885744 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51"} err="failed to get container status \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": rpc error: code = NotFound desc = could not find container \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": container with ID starting with 6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.885775 4679 scope.go:117] "RemoveContainer" containerID="40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.886123 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810"} err="failed to get container status \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": rpc error: code = NotFound desc = could not find container \"40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810\": container with ID starting with 40325580774e380495aab8d52706c33fdfe5ce8986586df927cc3658bfd01810 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.886150 4679 scope.go:117] "RemoveContainer" containerID="8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.886495 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799"} err="failed to get container status \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": rpc error: code = NotFound desc = could not find container \"8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799\": container with ID starting with 8fde1ae48efe0f0e52f09583ce9272e14d0942c8b4c8ec8b3f31214814150799 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.886523 4679 scope.go:117] "RemoveContainer" containerID="5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.887570 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf"} err="failed to get container status \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": rpc error: code = NotFound desc = could not find container \"5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf\": container with ID starting with 5f1dd266d59dd6f7c5b883faf2234c8544a9ff874eb2c89b892fc90e77ce58bf not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.887599 4679 scope.go:117] "RemoveContainer" containerID="6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.887946 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51"} err="failed to get container status \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": rpc error: code = NotFound desc = could not find container \"6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51\": container with ID starting with 6b3c036ae4c3caa4166a22cc1765f9f24882347b47d062764afa508acfc21c51 not found: ID does not exist" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.920585 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc553c8-18cf-4d05-a449-7ac987c8eb5d" path="/var/lib/kubelet/pods/7fc553c8-18cf-4d05-a449-7ac987c8eb5d/volumes" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.925787 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-config-data\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.925864 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-scripts\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.926031 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.926162 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8stdz\" (UniqueName: \"kubernetes.io/projected/ce2ce724-3129-47f2-a47a-c73e9f5576cd-kube-api-access-8stdz\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.926209 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-log-httpd\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.926384 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-run-httpd\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:10 crc kubenswrapper[4679]: I1202 10:40:10.926469 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.027651 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.027732 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-config-data\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.027823 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-scripts\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.027889 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.027944 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8stdz\" (UniqueName: \"kubernetes.io/projected/ce2ce724-3129-47f2-a47a-c73e9f5576cd-kube-api-access-8stdz\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.027967 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-log-httpd\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.028004 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-run-httpd\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.028329 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-run-httpd\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.029156 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-log-httpd\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.031564 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-scripts\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.032668 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.032678 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-config-data\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.032925 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.044885 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8stdz\" (UniqueName: \"kubernetes.io/projected/ce2ce724-3129-47f2-a47a-c73e9f5576cd-kube-api-access-8stdz\") pod \"ceilometer-0\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.166655 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.624664 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:11 crc kubenswrapper[4679]: W1202 10:40:11.631873 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce2ce724_3129_47f2_a47a_c73e9f5576cd.slice/crio-0739173764fa90fdf3b0627ff8d9eea8789f8c5dc1105fecc681fb941812e60f WatchSource:0}: Error finding container 0739173764fa90fdf3b0627ff8d9eea8789f8c5dc1105fecc681fb941812e60f: Status 404 returned error can't find the container with id 0739173764fa90fdf3b0627ff8d9eea8789f8c5dc1105fecc681fb941812e60f Dec 02 10:40:11 crc kubenswrapper[4679]: I1202 10:40:11.700017 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerStarted","Data":"0739173764fa90fdf3b0627ff8d9eea8789f8c5dc1105fecc681fb941812e60f"} Dec 02 10:40:16 crc kubenswrapper[4679]: I1202 10:40:16.932811 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:40:16 crc kubenswrapper[4679]: I1202 10:40:16.933406 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:40:16 crc kubenswrapper[4679]: I1202 10:40:16.933456 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:40:16 crc kubenswrapper[4679]: I1202 10:40:16.934582 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c09445f76d33fdda8f22af126b3018e23a8b609e3131d69d4172f788feafb09"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:40:16 crc kubenswrapper[4679]: I1202 10:40:16.934696 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://1c09445f76d33fdda8f22af126b3018e23a8b609e3131d69d4172f788feafb09" gracePeriod=600 Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.759443 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerStarted","Data":"1d8ba8e5aa030cca61f76f05247e5c15824e6017e04f7cd9a5540679004eb17d"} Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.761238 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" event={"ID":"b287fe6f-16df-4020-85b8-21ebdc224103","Type":"ContainerStarted","Data":"a053164b30b3974286b3971af458f15d99bfbffdcfe0513f8395aa141c0e7d26"} Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.766014 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="1c09445f76d33fdda8f22af126b3018e23a8b609e3131d69d4172f788feafb09" exitCode=0 Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.766061 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"1c09445f76d33fdda8f22af126b3018e23a8b609e3131d69d4172f788feafb09"} Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.766094 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"0891ad67c232e0d01b60d8bf4be071ef0e56a61b79e4a4f52127a08c972b2d9e"} Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.766110 4679 scope.go:117] "RemoveContainer" containerID="d0ace29b98cd54337ce3c65db2564b1430008dec833a4152e92f6248a82bf4d3" Dec 02 10:40:17 crc kubenswrapper[4679]: I1202 10:40:17.787698 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" podStartSLOduration=1.9498200350000001 podStartE2EDuration="9.787672224s" podCreationTimestamp="2025-12-02 10:40:08 +0000 UTC" firstStartedPulling="2025-12-02 10:40:09.265025569 +0000 UTC m=+1262.595164429" lastFinishedPulling="2025-12-02 10:40:17.102877758 +0000 UTC m=+1270.433016618" observedRunningTime="2025-12-02 10:40:17.780051554 +0000 UTC m=+1271.110190454" watchObservedRunningTime="2025-12-02 10:40:17.787672224 +0000 UTC m=+1271.117811134" Dec 02 10:40:18 crc kubenswrapper[4679]: I1202 10:40:18.777579 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerStarted","Data":"e8ec769056d0a018244b7c3776f0ce7019c0e08a27bb5bb230aed17faccebd1a"} Dec 02 10:40:19 crc kubenswrapper[4679]: I1202 10:40:19.793140 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerStarted","Data":"e84ae32bcfe5a9158e3f189a9abaf5c62f03f679d72ef1657aa70057953f2f68"} Dec 02 10:40:20 crc kubenswrapper[4679]: I1202 10:40:20.806984 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerStarted","Data":"c2dc0e5e1cbf278d9e3bb348987e252d165a74f68d43b1422d0c5ae6d93900e8"} Dec 02 10:40:20 crc kubenswrapper[4679]: I1202 10:40:20.808676 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:40:20 crc kubenswrapper[4679]: I1202 10:40:20.836543 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.033481125 podStartE2EDuration="10.836519798s" podCreationTimestamp="2025-12-02 10:40:10 +0000 UTC" firstStartedPulling="2025-12-02 10:40:11.63537831 +0000 UTC m=+1264.965517170" lastFinishedPulling="2025-12-02 10:40:20.438416953 +0000 UTC m=+1273.768555843" observedRunningTime="2025-12-02 10:40:20.825748947 +0000 UTC m=+1274.155887847" watchObservedRunningTime="2025-12-02 10:40:20.836519798 +0000 UTC m=+1274.166658688" Dec 02 10:40:30 crc kubenswrapper[4679]: I1202 10:40:30.962035 4679 generic.go:334] "Generic (PLEG): container finished" podID="b287fe6f-16df-4020-85b8-21ebdc224103" containerID="a053164b30b3974286b3971af458f15d99bfbffdcfe0513f8395aa141c0e7d26" exitCode=0 Dec 02 10:40:30 crc kubenswrapper[4679]: I1202 10:40:30.962181 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" event={"ID":"b287fe6f-16df-4020-85b8-21ebdc224103","Type":"ContainerDied","Data":"a053164b30b3974286b3971af458f15d99bfbffdcfe0513f8395aa141c0e7d26"} Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.293958 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.421764 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpkhx\" (UniqueName: \"kubernetes.io/projected/b287fe6f-16df-4020-85b8-21ebdc224103-kube-api-access-qpkhx\") pod \"b287fe6f-16df-4020-85b8-21ebdc224103\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.421911 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-scripts\") pod \"b287fe6f-16df-4020-85b8-21ebdc224103\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.421978 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-config-data\") pod \"b287fe6f-16df-4020-85b8-21ebdc224103\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.422035 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-combined-ca-bundle\") pod \"b287fe6f-16df-4020-85b8-21ebdc224103\" (UID: \"b287fe6f-16df-4020-85b8-21ebdc224103\") " Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.427929 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-scripts" (OuterVolumeSpecName: "scripts") pod "b287fe6f-16df-4020-85b8-21ebdc224103" (UID: "b287fe6f-16df-4020-85b8-21ebdc224103"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.428077 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b287fe6f-16df-4020-85b8-21ebdc224103-kube-api-access-qpkhx" (OuterVolumeSpecName: "kube-api-access-qpkhx") pod "b287fe6f-16df-4020-85b8-21ebdc224103" (UID: "b287fe6f-16df-4020-85b8-21ebdc224103"). InnerVolumeSpecName "kube-api-access-qpkhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.455004 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-config-data" (OuterVolumeSpecName: "config-data") pod "b287fe6f-16df-4020-85b8-21ebdc224103" (UID: "b287fe6f-16df-4020-85b8-21ebdc224103"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.455626 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b287fe6f-16df-4020-85b8-21ebdc224103" (UID: "b287fe6f-16df-4020-85b8-21ebdc224103"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.524847 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.524890 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.524905 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpkhx\" (UniqueName: \"kubernetes.io/projected/b287fe6f-16df-4020-85b8-21ebdc224103-kube-api-access-qpkhx\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.524918 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b287fe6f-16df-4020-85b8-21ebdc224103-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.979942 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" event={"ID":"b287fe6f-16df-4020-85b8-21ebdc224103","Type":"ContainerDied","Data":"dd3282054b10d3a916d64b92c0f67a0717bbd653c8b6fb3b4fb16b3cd717847a"} Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.979985 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3282054b10d3a916d64b92c0f67a0717bbd653c8b6fb3b4fb16b3cd717847a" Dec 02 10:40:32 crc kubenswrapper[4679]: I1202 10:40:32.980045 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bm8fd" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.094471 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 10:40:33 crc kubenswrapper[4679]: E1202 10:40:33.094864 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b287fe6f-16df-4020-85b8-21ebdc224103" containerName="nova-cell0-conductor-db-sync" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.094882 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="b287fe6f-16df-4020-85b8-21ebdc224103" containerName="nova-cell0-conductor-db-sync" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.095060 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="b287fe6f-16df-4020-85b8-21ebdc224103" containerName="nova-cell0-conductor-db-sync" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.095641 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.098859 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.099013 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jr68j" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.103965 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.238787 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.238963 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbcpv\" (UniqueName: \"kubernetes.io/projected/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-kube-api-access-qbcpv\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.239015 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.341436 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbcpv\" (UniqueName: \"kubernetes.io/projected/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-kube-api-access-qbcpv\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.341611 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.341771 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.345822 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.347068 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.361895 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbcpv\" (UniqueName: \"kubernetes.io/projected/759b2c49-d8b4-4f12-b60d-3ad7765f8aa3-kube-api-access-qbcpv\") pod \"nova-cell0-conductor-0\" (UID: \"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3\") " pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.422820 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.892720 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 10:40:33 crc kubenswrapper[4679]: W1202 10:40:33.893843 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod759b2c49_d8b4_4f12_b60d_3ad7765f8aa3.slice/crio-84e0de6377c58c8546531fdb0ef27ea49abc1e990fa31117dbea3a2e739cbe1b WatchSource:0}: Error finding container 84e0de6377c58c8546531fdb0ef27ea49abc1e990fa31117dbea3a2e739cbe1b: Status 404 returned error can't find the container with id 84e0de6377c58c8546531fdb0ef27ea49abc1e990fa31117dbea3a2e739cbe1b Dec 02 10:40:33 crc kubenswrapper[4679]: I1202 10:40:33.991776 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3","Type":"ContainerStarted","Data":"84e0de6377c58c8546531fdb0ef27ea49abc1e990fa31117dbea3a2e739cbe1b"} Dec 02 10:40:35 crc kubenswrapper[4679]: I1202 10:40:35.004300 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"759b2c49-d8b4-4f12-b60d-3ad7765f8aa3","Type":"ContainerStarted","Data":"c497a25f837b87892e55c20b00a173026700d51720bfaa08f0db884a38eb559b"} Dec 02 10:40:35 crc kubenswrapper[4679]: I1202 10:40:35.007235 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:35 crc kubenswrapper[4679]: I1202 10:40:35.027820 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.027799533 podStartE2EDuration="2.027799533s" podCreationTimestamp="2025-12-02 10:40:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:40:35.022594223 +0000 UTC m=+1288.352733153" watchObservedRunningTime="2025-12-02 10:40:35.027799533 +0000 UTC m=+1288.357938383" Dec 02 10:40:41 crc kubenswrapper[4679]: I1202 10:40:41.175245 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 02 10:40:43 crc kubenswrapper[4679]: I1202 10:40:43.450924 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 02 10:40:43 crc kubenswrapper[4679]: I1202 10:40:43.974345 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ttsxh"] Dec 02 10:40:43 crc kubenswrapper[4679]: I1202 10:40:43.978377 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:43 crc kubenswrapper[4679]: I1202 10:40:43.985144 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 02 10:40:43 crc kubenswrapper[4679]: I1202 10:40:43.985392 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 02 10:40:43 crc kubenswrapper[4679]: I1202 10:40:43.989125 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ttsxh"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.145233 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.146557 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.154090 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.154486 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpv8s\" (UniqueName: \"kubernetes.io/projected/7abb7421-f56a-462b-baea-b0030a020bd5-kube-api-access-bpv8s\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.154606 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-scripts\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.154629 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-config-data\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.154670 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.155833 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.221215 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.224134 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.226646 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.231955 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.233097 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.235314 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.247566 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263372 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-config-data\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263471 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-scripts\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263499 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-config-data\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263586 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263674 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263769 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpv8s\" (UniqueName: \"kubernetes.io/projected/7abb7421-f56a-462b-baea-b0030a020bd5-kube-api-access-bpv8s\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.263806 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnbb4\" (UniqueName: \"kubernetes.io/projected/dee36d9b-8726-49b5-b9d2-bb695c3442c3-kube-api-access-bnbb4\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.268130 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.280918 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-config-data\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.287422 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.302666 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-scripts\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.317256 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpv8s\" (UniqueName: \"kubernetes.io/projected/7abb7421-f56a-462b-baea-b0030a020bd5-kube-api-access-bpv8s\") pod \"nova-cell0-cell-mapping-ttsxh\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.323105 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.361405 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.363339 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.364970 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-config-data\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365021 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365047 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365086 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365111 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfrb\" (UniqueName: \"kubernetes.io/projected/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-kube-api-access-ppfrb\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365158 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365208 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7knr\" (UniqueName: \"kubernetes.io/projected/d17126b2-501f-4d23-96fa-91d694341f9b-kube-api-access-w7knr\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365233 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17126b2-501f-4d23-96fa-91d694341f9b-logs\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365252 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-config-data\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.365280 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnbb4\" (UniqueName: \"kubernetes.io/projected/dee36d9b-8726-49b5-b9d2-bb695c3442c3-kube-api-access-bnbb4\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.368361 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.398109 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.408601 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-config-data\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.413067 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnbb4\" (UniqueName: \"kubernetes.io/projected/dee36d9b-8726-49b5-b9d2-bb695c3442c3-kube-api-access-bnbb4\") pod \"nova-scheduler-0\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.415821 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.437372 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-drnpf"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.438863 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.443873 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-drnpf"] Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.475781 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483358 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483415 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfrb\" (UniqueName: \"kubernetes.io/projected/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-kube-api-access-ppfrb\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483509 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7knr\" (UniqueName: \"kubernetes.io/projected/d17126b2-501f-4d23-96fa-91d694341f9b-kube-api-access-w7knr\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483537 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17126b2-501f-4d23-96fa-91d694341f9b-logs\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483562 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-config-data\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483588 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q25kb\" (UniqueName: \"kubernetes.io/projected/b33e8170-003b-4405-9ffd-6d3161931ae7-kube-api-access-q25kb\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483627 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483655 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-config-data\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483730 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b33e8170-003b-4405-9ffd-6d3161931ae7-logs\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483797 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.483823 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.487856 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17126b2-501f-4d23-96fa-91d694341f9b-logs\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.493529 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.493992 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-config-data\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.498042 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.513155 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfrb\" (UniqueName: \"kubernetes.io/projected/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-kube-api-access-ppfrb\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.514685 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.547894 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7knr\" (UniqueName: \"kubernetes.io/projected/d17126b2-501f-4d23-96fa-91d694341f9b-kube-api-access-w7knr\") pod \"nova-api-0\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.562742 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.563807 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.585928 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.585977 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-config\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586037 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586071 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586098 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5bx4\" (UniqueName: \"kubernetes.io/projected/5d16bd4a-4e68-43d9-8b62-98825176cd75-kube-api-access-n5bx4\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586126 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q25kb\" (UniqueName: \"kubernetes.io/projected/b33e8170-003b-4405-9ffd-6d3161931ae7-kube-api-access-q25kb\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586141 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586171 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586189 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-config-data\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586259 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b33e8170-003b-4405-9ffd-6d3161931ae7-logs\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.586700 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b33e8170-003b-4405-9ffd-6d3161931ae7-logs\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.591748 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.599721 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-config-data\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.607768 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q25kb\" (UniqueName: \"kubernetes.io/projected/b33e8170-003b-4405-9ffd-6d3161931ae7-kube-api-access-q25kb\") pod \"nova-metadata-0\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.689323 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.689417 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-config\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.690361 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-config\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.690400 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.690510 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.690611 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.690677 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5bx4\" (UniqueName: \"kubernetes.io/projected/5d16bd4a-4e68-43d9-8b62-98825176cd75-kube-api-access-n5bx4\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.690716 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.691204 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.691870 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.692129 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.714414 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5bx4\" (UniqueName: \"kubernetes.io/projected/5d16bd4a-4e68-43d9-8b62-98825176cd75-kube-api-access-n5bx4\") pod \"dnsmasq-dns-845d6d6f59-drnpf\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.828002 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.840021 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:44 crc kubenswrapper[4679]: I1202 10:40:44.932516 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ttsxh"] Dec 02 10:40:44 crc kubenswrapper[4679]: W1202 10:40:44.940942 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7abb7421_f56a_462b_baea_b0030a020bd5.slice/crio-013d3aa2c980a155aca61e3d9766639c220f28153ab5d8a455be4a14a9f92e78 WatchSource:0}: Error finding container 013d3aa2c980a155aca61e3d9766639c220f28153ab5d8a455be4a14a9f92e78: Status 404 returned error can't find the container with id 013d3aa2c980a155aca61e3d9766639c220f28153ab5d8a455be4a14a9f92e78 Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.062162 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-md655"] Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.065275 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.067943 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.068183 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.086930 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-md655"] Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.104332 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.122830 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dee36d9b-8726-49b5-b9d2-bb695c3442c3","Type":"ContainerStarted","Data":"4b0974d3052b6ce3c8e36e4cf1d63d5e954f86e8065b308a6e308a25b0c1f9da"} Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.124789 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ttsxh" event={"ID":"7abb7421-f56a-462b-baea-b0030a020bd5","Type":"ContainerStarted","Data":"013d3aa2c980a155aca61e3d9766639c220f28153ab5d8a455be4a14a9f92e78"} Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.202825 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8khg8\" (UniqueName: \"kubernetes.io/projected/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-kube-api-access-8khg8\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.203506 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-scripts\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.203807 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-config-data\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.204090 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.206188 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.305598 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8khg8\" (UniqueName: \"kubernetes.io/projected/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-kube-api-access-8khg8\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.305646 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-scripts\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.305675 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-config-data\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.305773 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.312771 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-scripts\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.312847 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.313546 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-config-data\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.333546 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8khg8\" (UniqueName: \"kubernetes.io/projected/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-kube-api-access-8khg8\") pod \"nova-cell1-conductor-db-sync-md655\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.372222 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.394868 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.490159 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-drnpf"] Dec 02 10:40:45 crc kubenswrapper[4679]: W1202 10:40:45.494578 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb33e8170_003b_4405_9ffd_6d3161931ae7.slice/crio-1805aa6492a5292a01dfd75620e2b8cfa1fc6a179e6254d95e604c2e391cb41c WatchSource:0}: Error finding container 1805aa6492a5292a01dfd75620e2b8cfa1fc6a179e6254d95e604c2e391cb41c: Status 404 returned error can't find the container with id 1805aa6492a5292a01dfd75620e2b8cfa1fc6a179e6254d95e604c2e391cb41c Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.504700 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:45 crc kubenswrapper[4679]: I1202 10:40:45.918559 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-md655"] Dec 02 10:40:45 crc kubenswrapper[4679]: W1202 10:40:45.921459 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c5b0d9f_8c3f_478e_9622_44513c5bd88c.slice/crio-e091282b4fc6c32907f8d9359b35ae3084776045ef6ef570d541e462ceec9842 WatchSource:0}: Error finding container e091282b4fc6c32907f8d9359b35ae3084776045ef6ef570d541e462ceec9842: Status 404 returned error can't find the container with id e091282b4fc6c32907f8d9359b35ae3084776045ef6ef570d541e462ceec9842 Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.140814 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-md655" event={"ID":"7c5b0d9f-8c3f-478e-9622-44513c5bd88c","Type":"ContainerStarted","Data":"e091282b4fc6c32907f8d9359b35ae3084776045ef6ef570d541e462ceec9842"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.145104 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d","Type":"ContainerStarted","Data":"f8447faa31284d856fee03e0a1dcccf22679fd4add6ac2c479e0e36c6d5d1a85"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.146809 4679 generic.go:334] "Generic (PLEG): container finished" podID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerID="39f33088a3a25c2c1cd3e8c91ee418ea04d4016736a7296fe08ccc7b241ed237" exitCode=0 Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.147084 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" event={"ID":"5d16bd4a-4e68-43d9-8b62-98825176cd75","Type":"ContainerDied","Data":"39f33088a3a25c2c1cd3e8c91ee418ea04d4016736a7296fe08ccc7b241ed237"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.147124 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" event={"ID":"5d16bd4a-4e68-43d9-8b62-98825176cd75","Type":"ContainerStarted","Data":"ef13fcb044157e7d8ec6777ca2bd382261bfc6365949e7d3bebc67c5e3c719fe"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.150232 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ttsxh" event={"ID":"7abb7421-f56a-462b-baea-b0030a020bd5","Type":"ContainerStarted","Data":"184efdb137bd8381f06c71a3d4fc9e363a201b001cbb34d0699471c10b2dcab2"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.154035 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b33e8170-003b-4405-9ffd-6d3161931ae7","Type":"ContainerStarted","Data":"1805aa6492a5292a01dfd75620e2b8cfa1fc6a179e6254d95e604c2e391cb41c"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.160060 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d17126b2-501f-4d23-96fa-91d694341f9b","Type":"ContainerStarted","Data":"1140491dd2e3a1df97f0c05f5bf4eea61305f43fd6faf541b4bb0a62592b5070"} Dec 02 10:40:46 crc kubenswrapper[4679]: I1202 10:40:46.191733 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ttsxh" podStartSLOduration=3.191713656 podStartE2EDuration="3.191713656s" podCreationTimestamp="2025-12-02 10:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:40:46.183177701 +0000 UTC m=+1299.513316581" watchObservedRunningTime="2025-12-02 10:40:46.191713656 +0000 UTC m=+1299.521852516" Dec 02 10:40:47 crc kubenswrapper[4679]: I1202 10:40:47.207223 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-md655" event={"ID":"7c5b0d9f-8c3f-478e-9622-44513c5bd88c","Type":"ContainerStarted","Data":"a8bd6dae67bddf4f4f5d13f500e2c7e7950a36902cbcb8e667aa14132735a428"} Dec 02 10:40:47 crc kubenswrapper[4679]: I1202 10:40:47.227678 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dee36d9b-8726-49b5-b9d2-bb695c3442c3","Type":"ContainerStarted","Data":"caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0"} Dec 02 10:40:47 crc kubenswrapper[4679]: I1202 10:40:47.246936 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" event={"ID":"5d16bd4a-4e68-43d9-8b62-98825176cd75","Type":"ContainerStarted","Data":"adc0b6d07a9e004ccd5957f680457754553a5174e4b24399f291e77d4c49a63b"} Dec 02 10:40:47 crc kubenswrapper[4679]: I1202 10:40:47.248451 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-md655" podStartSLOduration=3.248432323 podStartE2EDuration="3.248432323s" podCreationTimestamp="2025-12-02 10:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:40:47.237758956 +0000 UTC m=+1300.567897816" watchObservedRunningTime="2025-12-02 10:40:47.248432323 +0000 UTC m=+1300.578571183" Dec 02 10:40:47 crc kubenswrapper[4679]: I1202 10:40:47.281533 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.77582034 podStartE2EDuration="3.281509894s" podCreationTimestamp="2025-12-02 10:40:44 +0000 UTC" firstStartedPulling="2025-12-02 10:40:45.097515992 +0000 UTC m=+1298.427654852" lastFinishedPulling="2025-12-02 10:40:46.603205536 +0000 UTC m=+1299.933344406" observedRunningTime="2025-12-02 10:40:47.277869819 +0000 UTC m=+1300.608008679" watchObservedRunningTime="2025-12-02 10:40:47.281509894 +0000 UTC m=+1300.611648774" Dec 02 10:40:47 crc kubenswrapper[4679]: I1202 10:40:47.313206 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" podStartSLOduration=3.313182964 podStartE2EDuration="3.313182964s" podCreationTimestamp="2025-12-02 10:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:40:47.309382235 +0000 UTC m=+1300.639521115" watchObservedRunningTime="2025-12-02 10:40:47.313182964 +0000 UTC m=+1300.643321824" Dec 02 10:40:48 crc kubenswrapper[4679]: I1202 10:40:48.257196 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.080754 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.094637 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.203801 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.204207 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e2a75d52-ee01-460e-b15b-70bff6593ef1" containerName="kube-state-metrics" containerID="cri-o://44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b" gracePeriod=30 Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.264794 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d","Type":"ContainerStarted","Data":"a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1"} Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.266340 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b33e8170-003b-4405-9ffd-6d3161931ae7","Type":"ContainerStarted","Data":"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f"} Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.266402 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b33e8170-003b-4405-9ffd-6d3161931ae7","Type":"ContainerStarted","Data":"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e"} Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.267881 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d17126b2-501f-4d23-96fa-91d694341f9b","Type":"ContainerStarted","Data":"b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df"} Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.267927 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d17126b2-501f-4d23-96fa-91d694341f9b","Type":"ContainerStarted","Data":"96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd"} Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.281507 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.335795948 podStartE2EDuration="5.281488597s" podCreationTimestamp="2025-12-02 10:40:44 +0000 UTC" firstStartedPulling="2025-12-02 10:40:45.379727604 +0000 UTC m=+1298.709866464" lastFinishedPulling="2025-12-02 10:40:48.325420253 +0000 UTC m=+1301.655559113" observedRunningTime="2025-12-02 10:40:49.278711707 +0000 UTC m=+1302.608850567" watchObservedRunningTime="2025-12-02 10:40:49.281488597 +0000 UTC m=+1302.611627457" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.304832 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.481219619 podStartE2EDuration="5.304810938s" podCreationTimestamp="2025-12-02 10:40:44 +0000 UTC" firstStartedPulling="2025-12-02 10:40:45.501867935 +0000 UTC m=+1298.832006795" lastFinishedPulling="2025-12-02 10:40:48.325459254 +0000 UTC m=+1301.655598114" observedRunningTime="2025-12-02 10:40:49.29829055 +0000 UTC m=+1302.628429410" watchObservedRunningTime="2025-12-02 10:40:49.304810938 +0000 UTC m=+1302.634949808" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.341615 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.230820221 podStartE2EDuration="5.341593375s" podCreationTimestamp="2025-12-02 10:40:44 +0000 UTC" firstStartedPulling="2025-12-02 10:40:45.21469613 +0000 UTC m=+1298.544834990" lastFinishedPulling="2025-12-02 10:40:48.325469284 +0000 UTC m=+1301.655608144" observedRunningTime="2025-12-02 10:40:49.317943435 +0000 UTC m=+1302.648082295" watchObservedRunningTime="2025-12-02 10:40:49.341593375 +0000 UTC m=+1302.671732225" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.476741 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.564644 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.742811 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.810744 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqtbj\" (UniqueName: \"kubernetes.io/projected/e2a75d52-ee01-460e-b15b-70bff6593ef1-kube-api-access-lqtbj\") pod \"e2a75d52-ee01-460e-b15b-70bff6593ef1\" (UID: \"e2a75d52-ee01-460e-b15b-70bff6593ef1\") " Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.817425 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a75d52-ee01-460e-b15b-70bff6593ef1-kube-api-access-lqtbj" (OuterVolumeSpecName: "kube-api-access-lqtbj") pod "e2a75d52-ee01-460e-b15b-70bff6593ef1" (UID: "e2a75d52-ee01-460e-b15b-70bff6593ef1"). InnerVolumeSpecName "kube-api-access-lqtbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.828181 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.828275 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:40:49 crc kubenswrapper[4679]: I1202 10:40:49.913461 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqtbj\" (UniqueName: \"kubernetes.io/projected/e2a75d52-ee01-460e-b15b-70bff6593ef1-kube-api-access-lqtbj\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.279728 4679 generic.go:334] "Generic (PLEG): container finished" podID="e2a75d52-ee01-460e-b15b-70bff6593ef1" containerID="44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b" exitCode=2 Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.279803 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.279871 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2a75d52-ee01-460e-b15b-70bff6593ef1","Type":"ContainerDied","Data":"44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b"} Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.279930 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2a75d52-ee01-460e-b15b-70bff6593ef1","Type":"ContainerDied","Data":"89819e526360acedcc60e7831a7b9c8ceccf084b4d96617cd33ee97cc05ebd6c"} Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.279954 4679 scope.go:117] "RemoveContainer" containerID="44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.280691 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-log" containerID="cri-o://28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e" gracePeriod=30 Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.280841 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1" gracePeriod=30 Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.280937 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-metadata" containerID="cri-o://fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f" gracePeriod=30 Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.318796 4679 scope.go:117] "RemoveContainer" containerID="44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b" Dec 02 10:40:50 crc kubenswrapper[4679]: E1202 10:40:50.323755 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b\": container with ID starting with 44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b not found: ID does not exist" containerID="44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.323805 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b"} err="failed to get container status \"44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b\": rpc error: code = NotFound desc = could not find container \"44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b\": container with ID starting with 44cdb03a5d6c858d8f6b4883002d0b38c3db9e6afdab977adc49078bc2111b9b not found: ID does not exist" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.331480 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.350410 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.373426 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:40:50 crc kubenswrapper[4679]: E1202 10:40:50.378033 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a75d52-ee01-460e-b15b-70bff6593ef1" containerName="kube-state-metrics" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.380510 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a75d52-ee01-460e-b15b-70bff6593ef1" containerName="kube-state-metrics" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.381617 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2a75d52-ee01-460e-b15b-70bff6593ef1" containerName="kube-state-metrics" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.383168 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.383357 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.386812 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.386873 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.524019 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.524091 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99mkh\" (UniqueName: \"kubernetes.io/projected/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-api-access-99mkh\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.524119 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.524141 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.626285 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.626889 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99mkh\" (UniqueName: \"kubernetes.io/projected/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-api-access-99mkh\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.626920 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.626946 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.633035 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.634318 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.649924 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99mkh\" (UniqueName: \"kubernetes.io/projected/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-api-access-99mkh\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.666959 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84bc340a-3e0b-4ac5-94a7-07156dbcbbf0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0\") " pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.751889 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.851054 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.930391 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2a75d52-ee01-460e-b15b-70bff6593ef1" path="/var/lib/kubelet/pods/e2a75d52-ee01-460e-b15b-70bff6593ef1/volumes" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.938294 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b33e8170-003b-4405-9ffd-6d3161931ae7-logs\") pod \"b33e8170-003b-4405-9ffd-6d3161931ae7\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.938579 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q25kb\" (UniqueName: \"kubernetes.io/projected/b33e8170-003b-4405-9ffd-6d3161931ae7-kube-api-access-q25kb\") pod \"b33e8170-003b-4405-9ffd-6d3161931ae7\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.938624 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33e8170-003b-4405-9ffd-6d3161931ae7-logs" (OuterVolumeSpecName: "logs") pod "b33e8170-003b-4405-9ffd-6d3161931ae7" (UID: "b33e8170-003b-4405-9ffd-6d3161931ae7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.938691 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-combined-ca-bundle\") pod \"b33e8170-003b-4405-9ffd-6d3161931ae7\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.938896 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-config-data\") pod \"b33e8170-003b-4405-9ffd-6d3161931ae7\" (UID: \"b33e8170-003b-4405-9ffd-6d3161931ae7\") " Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.939684 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b33e8170-003b-4405-9ffd-6d3161931ae7-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.966457 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33e8170-003b-4405-9ffd-6d3161931ae7-kube-api-access-q25kb" (OuterVolumeSpecName: "kube-api-access-q25kb") pod "b33e8170-003b-4405-9ffd-6d3161931ae7" (UID: "b33e8170-003b-4405-9ffd-6d3161931ae7"). InnerVolumeSpecName "kube-api-access-q25kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:50 crc kubenswrapper[4679]: I1202 10:40:50.987621 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-config-data" (OuterVolumeSpecName: "config-data") pod "b33e8170-003b-4405-9ffd-6d3161931ae7" (UID: "b33e8170-003b-4405-9ffd-6d3161931ae7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.001378 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b33e8170-003b-4405-9ffd-6d3161931ae7" (UID: "b33e8170-003b-4405-9ffd-6d3161931ae7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.042453 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q25kb\" (UniqueName: \"kubernetes.io/projected/b33e8170-003b-4405-9ffd-6d3161931ae7-kube-api-access-q25kb\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.042481 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.042496 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b33e8170-003b-4405-9ffd-6d3161931ae7-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:51 crc kubenswrapper[4679]: W1202 10:40:51.259282 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84bc340a_3e0b_4ac5_94a7_07156dbcbbf0.slice/crio-df4508e5433bb88fd8058bbf91e09acaee59dcab584291d3c5d9eefeb3d82cf5 WatchSource:0}: Error finding container df4508e5433bb88fd8058bbf91e09acaee59dcab584291d3c5d9eefeb3d82cf5: Status 404 returned error can't find the container with id df4508e5433bb88fd8058bbf91e09acaee59dcab584291d3c5d9eefeb3d82cf5 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.259465 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.289910 4679 generic.go:334] "Generic (PLEG): container finished" podID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerID="fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f" exitCode=0 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.289950 4679 generic.go:334] "Generic (PLEG): container finished" podID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerID="28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e" exitCode=143 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.289963 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.290038 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b33e8170-003b-4405-9ffd-6d3161931ae7","Type":"ContainerDied","Data":"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f"} Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.290123 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b33e8170-003b-4405-9ffd-6d3161931ae7","Type":"ContainerDied","Data":"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e"} Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.290141 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b33e8170-003b-4405-9ffd-6d3161931ae7","Type":"ContainerDied","Data":"1805aa6492a5292a01dfd75620e2b8cfa1fc6a179e6254d95e604c2e391cb41c"} Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.290164 4679 scope.go:117] "RemoveContainer" containerID="fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.291623 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0","Type":"ContainerStarted","Data":"df4508e5433bb88fd8058bbf91e09acaee59dcab584291d3c5d9eefeb3d82cf5"} Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.317393 4679 scope.go:117] "RemoveContainer" containerID="28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.321967 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.332365 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.346379 4679 scope.go:117] "RemoveContainer" containerID="fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f" Dec 02 10:40:51 crc kubenswrapper[4679]: E1202 10:40:51.348841 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f\": container with ID starting with fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f not found: ID does not exist" containerID="fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.348887 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f"} err="failed to get container status \"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f\": rpc error: code = NotFound desc = could not find container \"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f\": container with ID starting with fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f not found: ID does not exist" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.348911 4679 scope.go:117] "RemoveContainer" containerID="28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e" Dec 02 10:40:51 crc kubenswrapper[4679]: E1202 10:40:51.349281 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e\": container with ID starting with 28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e not found: ID does not exist" containerID="28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.349320 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e"} err="failed to get container status \"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e\": rpc error: code = NotFound desc = could not find container \"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e\": container with ID starting with 28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e not found: ID does not exist" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.349345 4679 scope.go:117] "RemoveContainer" containerID="fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.349574 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.349706 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f"} err="failed to get container status \"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f\": rpc error: code = NotFound desc = could not find container \"fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f\": container with ID starting with fd69174146a935d663239c99865d20ab75330eb01c55a3a1ee0e4c16406e758f not found: ID does not exist" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.349748 4679 scope.go:117] "RemoveContainer" containerID="28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e" Dec 02 10:40:51 crc kubenswrapper[4679]: E1202 10:40:51.350044 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-log" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.350069 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-log" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.350069 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e"} err="failed to get container status \"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e\": rpc error: code = NotFound desc = could not find container \"28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e\": container with ID starting with 28d2fb3b6d45f49ef3dfe85b1bbf404266399f483f194118906ff0cf68a4073e not found: ID does not exist" Dec 02 10:40:51 crc kubenswrapper[4679]: E1202 10:40:51.350105 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-metadata" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.350118 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-metadata" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.350380 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-metadata" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.350404 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" containerName="nova-metadata-log" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.351343 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.359839 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.360036 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.360690 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.448223 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf5kp\" (UniqueName: \"kubernetes.io/projected/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-kube-api-access-vf5kp\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.448274 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-config-data\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.448379 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-logs\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.448473 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.448496 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.550445 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.550828 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.550933 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf5kp\" (UniqueName: \"kubernetes.io/projected/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-kube-api-access-vf5kp\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.550960 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-config-data\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.550997 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-logs\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.551404 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-logs\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.556080 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.558106 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.559718 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-config-data\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.569822 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf5kp\" (UniqueName: \"kubernetes.io/projected/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-kube-api-access-vf5kp\") pod \"nova-metadata-0\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " pod="openstack/nova-metadata-0" Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.616433 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.616986 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-central-agent" containerID="cri-o://1d8ba8e5aa030cca61f76f05247e5c15824e6017e04f7cd9a5540679004eb17d" gracePeriod=30 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.617084 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="sg-core" containerID="cri-o://e84ae32bcfe5a9158e3f189a9abaf5c62f03f679d72ef1657aa70057953f2f68" gracePeriod=30 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.617259 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-notification-agent" containerID="cri-o://e8ec769056d0a018244b7c3776f0ce7019c0e08a27bb5bb230aed17faccebd1a" gracePeriod=30 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.617042 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="proxy-httpd" containerID="cri-o://c2dc0e5e1cbf278d9e3bb348987e252d165a74f68d43b1422d0c5ae6d93900e8" gracePeriod=30 Dec 02 10:40:51 crc kubenswrapper[4679]: I1202 10:40:51.671493 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.250015 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:52 crc kubenswrapper[4679]: W1202 10:40:52.254205 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13b59b5e_6c16_48e0_9e32_a4f4c9fd97fe.slice/crio-b6ab9cca5b6222eceddc98fb5bf46663fba3e83696db056993f7de75fdcadbd9 WatchSource:0}: Error finding container b6ab9cca5b6222eceddc98fb5bf46663fba3e83696db056993f7de75fdcadbd9: Status 404 returned error can't find the container with id b6ab9cca5b6222eceddc98fb5bf46663fba3e83696db056993f7de75fdcadbd9 Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.302246 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84bc340a-3e0b-4ac5-94a7-07156dbcbbf0","Type":"ContainerStarted","Data":"c3b8173517437f4b22a8b19a5ed14978f65af04c187291f43d701408f7e4e39e"} Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.303401 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.307301 4679 generic.go:334] "Generic (PLEG): container finished" podID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerID="c2dc0e5e1cbf278d9e3bb348987e252d165a74f68d43b1422d0c5ae6d93900e8" exitCode=0 Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.307410 4679 generic.go:334] "Generic (PLEG): container finished" podID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerID="e84ae32bcfe5a9158e3f189a9abaf5c62f03f679d72ef1657aa70057953f2f68" exitCode=2 Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.307420 4679 generic.go:334] "Generic (PLEG): container finished" podID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerID="1d8ba8e5aa030cca61f76f05247e5c15824e6017e04f7cd9a5540679004eb17d" exitCode=0 Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.307452 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerDied","Data":"c2dc0e5e1cbf278d9e3bb348987e252d165a74f68d43b1422d0c5ae6d93900e8"} Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.307470 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerDied","Data":"e84ae32bcfe5a9158e3f189a9abaf5c62f03f679d72ef1657aa70057953f2f68"} Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.307480 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerDied","Data":"1d8ba8e5aa030cca61f76f05247e5c15824e6017e04f7cd9a5540679004eb17d"} Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.309487 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe","Type":"ContainerStarted","Data":"b6ab9cca5b6222eceddc98fb5bf46663fba3e83696db056993f7de75fdcadbd9"} Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.323215 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9299406110000001 podStartE2EDuration="2.323195386s" podCreationTimestamp="2025-12-02 10:40:50 +0000 UTC" firstStartedPulling="2025-12-02 10:40:51.261587878 +0000 UTC m=+1304.591726738" lastFinishedPulling="2025-12-02 10:40:51.654842653 +0000 UTC m=+1304.984981513" observedRunningTime="2025-12-02 10:40:52.322545627 +0000 UTC m=+1305.652684487" watchObservedRunningTime="2025-12-02 10:40:52.323195386 +0000 UTC m=+1305.653334246" Dec 02 10:40:52 crc kubenswrapper[4679]: I1202 10:40:52.923901 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33e8170-003b-4405-9ffd-6d3161931ae7" path="/var/lib/kubelet/pods/b33e8170-003b-4405-9ffd-6d3161931ae7/volumes" Dec 02 10:40:53 crc kubenswrapper[4679]: I1202 10:40:53.334653 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe","Type":"ContainerStarted","Data":"09a5370e9d2bb3ccead6f3ab1604562f5c2d5842e808a189eba8836efb7f28d0"} Dec 02 10:40:53 crc kubenswrapper[4679]: I1202 10:40:53.335125 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe","Type":"ContainerStarted","Data":"51fbfe9ca51bb63f94d28d8e629804c8b943dbeee498ec2291f9894d9975073c"} Dec 02 10:40:53 crc kubenswrapper[4679]: I1202 10:40:53.366574 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.366538909 podStartE2EDuration="2.366538909s" podCreationTimestamp="2025-12-02 10:40:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:40:53.359716783 +0000 UTC m=+1306.689855643" watchObservedRunningTime="2025-12-02 10:40:53.366538909 +0000 UTC m=+1306.696677769" Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.362006 4679 generic.go:334] "Generic (PLEG): container finished" podID="7abb7421-f56a-462b-baea-b0030a020bd5" containerID="184efdb137bd8381f06c71a3d4fc9e363a201b001cbb34d0699471c10b2dcab2" exitCode=0 Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.362152 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ttsxh" event={"ID":"7abb7421-f56a-462b-baea-b0030a020bd5","Type":"ContainerDied","Data":"184efdb137bd8381f06c71a3d4fc9e363a201b001cbb34d0699471c10b2dcab2"} Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.365094 4679 generic.go:334] "Generic (PLEG): container finished" podID="7c5b0d9f-8c3f-478e-9622-44513c5bd88c" containerID="a8bd6dae67bddf4f4f5d13f500e2c7e7950a36902cbcb8e667aa14132735a428" exitCode=0 Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.365163 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-md655" event={"ID":"7c5b0d9f-8c3f-478e-9622-44513c5bd88c","Type":"ContainerDied","Data":"a8bd6dae67bddf4f4f5d13f500e2c7e7950a36902cbcb8e667aa14132735a428"} Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.476867 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.511158 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.564745 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.564799 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.841513 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.957744 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2wk9g"] Dec 02 10:40:54 crc kubenswrapper[4679]: I1202 10:40:54.958112 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerName="dnsmasq-dns" containerID="cri-o://0c77363366fe5b511f74f99f79a81bd694936505421107887cbd6649a574381c" gracePeriod=10 Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.377539 4679 generic.go:334] "Generic (PLEG): container finished" podID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerID="0c77363366fe5b511f74f99f79a81bd694936505421107887cbd6649a574381c" exitCode=0 Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.377586 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" event={"ID":"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d","Type":"ContainerDied","Data":"0c77363366fe5b511f74f99f79a81bd694936505421107887cbd6649a574381c"} Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.421249 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.649446 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.649852 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.826166 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:55 crc kubenswrapper[4679]: I1202 10:40:55.832013 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.009798 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-scripts\") pod \"7abb7421-f56a-462b-baea-b0030a020bd5\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.009852 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-combined-ca-bundle\") pod \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.009927 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8khg8\" (UniqueName: \"kubernetes.io/projected/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-kube-api-access-8khg8\") pod \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.010038 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-config-data\") pod \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.010076 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpv8s\" (UniqueName: \"kubernetes.io/projected/7abb7421-f56a-462b-baea-b0030a020bd5-kube-api-access-bpv8s\") pod \"7abb7421-f56a-462b-baea-b0030a020bd5\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.010096 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-config-data\") pod \"7abb7421-f56a-462b-baea-b0030a020bd5\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.010135 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-combined-ca-bundle\") pod \"7abb7421-f56a-462b-baea-b0030a020bd5\" (UID: \"7abb7421-f56a-462b-baea-b0030a020bd5\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.010169 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-scripts\") pod \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\" (UID: \"7c5b0d9f-8c3f-478e-9622-44513c5bd88c\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.016673 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-scripts" (OuterVolumeSpecName: "scripts") pod "7abb7421-f56a-462b-baea-b0030a020bd5" (UID: "7abb7421-f56a-462b-baea-b0030a020bd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.019473 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-scripts" (OuterVolumeSpecName: "scripts") pod "7c5b0d9f-8c3f-478e-9622-44513c5bd88c" (UID: "7c5b0d9f-8c3f-478e-9622-44513c5bd88c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.019526 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7abb7421-f56a-462b-baea-b0030a020bd5-kube-api-access-bpv8s" (OuterVolumeSpecName: "kube-api-access-bpv8s") pod "7abb7421-f56a-462b-baea-b0030a020bd5" (UID: "7abb7421-f56a-462b-baea-b0030a020bd5"). InnerVolumeSpecName "kube-api-access-bpv8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.029547 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-kube-api-access-8khg8" (OuterVolumeSpecName: "kube-api-access-8khg8") pod "7c5b0d9f-8c3f-478e-9622-44513c5bd88c" (UID: "7c5b0d9f-8c3f-478e-9622-44513c5bd88c"). InnerVolumeSpecName "kube-api-access-8khg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.053120 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c5b0d9f-8c3f-478e-9622-44513c5bd88c" (UID: "7c5b0d9f-8c3f-478e-9622-44513c5bd88c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.057433 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7abb7421-f56a-462b-baea-b0030a020bd5" (UID: "7abb7421-f56a-462b-baea-b0030a020bd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.061394 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-config-data" (OuterVolumeSpecName: "config-data") pod "7c5b0d9f-8c3f-478e-9622-44513c5bd88c" (UID: "7c5b0d9f-8c3f-478e-9622-44513c5bd88c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.061845 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.073797 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-config-data" (OuterVolumeSpecName: "config-data") pod "7abb7421-f56a-462b-baea-b0030a020bd5" (UID: "7abb7421-f56a-462b-baea-b0030a020bd5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112266 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpv8s\" (UniqueName: \"kubernetes.io/projected/7abb7421-f56a-462b-baea-b0030a020bd5-kube-api-access-bpv8s\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112340 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112352 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112361 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112372 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7abb7421-f56a-462b-baea-b0030a020bd5-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112380 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112388 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8khg8\" (UniqueName: \"kubernetes.io/projected/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-kube-api-access-8khg8\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.112416 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c5b0d9f-8c3f-478e-9622-44513c5bd88c-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.213406 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-config\") pod \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.213516 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-svc\") pod \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.213572 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds98j\" (UniqueName: \"kubernetes.io/projected/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-kube-api-access-ds98j\") pod \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.213633 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-nb\") pod \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.213706 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-sb\") pod \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.213751 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-swift-storage-0\") pod \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\" (UID: \"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d\") " Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.219932 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-kube-api-access-ds98j" (OuterVolumeSpecName: "kube-api-access-ds98j") pod "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" (UID: "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d"). InnerVolumeSpecName "kube-api-access-ds98j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.267995 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" (UID: "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.276668 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-config" (OuterVolumeSpecName: "config") pod "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" (UID: "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.277285 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" (UID: "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.282418 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" (UID: "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.295757 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" (UID: "4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.316557 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.316587 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.316602 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds98j\" (UniqueName: \"kubernetes.io/projected/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-kube-api-access-ds98j\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.316615 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.316627 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.316639 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.402059 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" event={"ID":"4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d","Type":"ContainerDied","Data":"2456db17fccc338f646e889cacf3875f21c504beeda0eea4d64550a9424f8c52"} Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.402433 4679 scope.go:117] "RemoveContainer" containerID="0c77363366fe5b511f74f99f79a81bd694936505421107887cbd6649a574381c" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.402734 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2wk9g" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.404038 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ttsxh" event={"ID":"7abb7421-f56a-462b-baea-b0030a020bd5","Type":"ContainerDied","Data":"013d3aa2c980a155aca61e3d9766639c220f28153ab5d8a455be4a14a9f92e78"} Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.404078 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="013d3aa2c980a155aca61e3d9766639c220f28153ab5d8a455be4a14a9f92e78" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.404140 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ttsxh" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.409368 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-md655" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.410030 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-md655" event={"ID":"7c5b0d9f-8c3f-478e-9622-44513c5bd88c","Type":"ContainerDied","Data":"e091282b4fc6c32907f8d9359b35ae3084776045ef6ef570d541e462ceec9842"} Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.410070 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e091282b4fc6c32907f8d9359b35ae3084776045ef6ef570d541e462ceec9842" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.453770 4679 scope.go:117] "RemoveContainer" containerID="b3c65d1c5056672272bb1f880561342fd3be89d0cb449d5f35a743fac78a3d33" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.475365 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2wk9g"] Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.500703 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2wk9g"] Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.521047 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 10:40:56 crc kubenswrapper[4679]: E1202 10:40:56.521751 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7abb7421-f56a-462b-baea-b0030a020bd5" containerName="nova-manage" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.521775 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7abb7421-f56a-462b-baea-b0030a020bd5" containerName="nova-manage" Dec 02 10:40:56 crc kubenswrapper[4679]: E1202 10:40:56.521800 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5b0d9f-8c3f-478e-9622-44513c5bd88c" containerName="nova-cell1-conductor-db-sync" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.521809 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5b0d9f-8c3f-478e-9622-44513c5bd88c" containerName="nova-cell1-conductor-db-sync" Dec 02 10:40:56 crc kubenswrapper[4679]: E1202 10:40:56.521826 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerName="dnsmasq-dns" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.521888 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerName="dnsmasq-dns" Dec 02 10:40:56 crc kubenswrapper[4679]: E1202 10:40:56.521930 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerName="init" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.521939 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerName="init" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.522246 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c5b0d9f-8c3f-478e-9622-44513c5bd88c" containerName="nova-cell1-conductor-db-sync" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.522280 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" containerName="dnsmasq-dns" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.522335 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7abb7421-f56a-462b-baea-b0030a020bd5" containerName="nova-manage" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.523493 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.525685 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.536747 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.601364 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.601770 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-api" containerID="cri-o://b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df" gracePeriod=30 Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.601728 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-log" containerID="cri-o://96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd" gracePeriod=30 Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.612465 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.623442 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95edebcc-4231-4f68-be9d-7e04a0d15e16-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.623502 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwx5g\" (UniqueName: \"kubernetes.io/projected/95edebcc-4231-4f68-be9d-7e04a0d15e16-kube-api-access-wwx5g\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.623639 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95edebcc-4231-4f68-be9d-7e04a0d15e16-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.656177 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.656466 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-log" containerID="cri-o://51fbfe9ca51bb63f94d28d8e629804c8b943dbeee498ec2291f9894d9975073c" gracePeriod=30 Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.656947 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-metadata" containerID="cri-o://09a5370e9d2bb3ccead6f3ab1604562f5c2d5842e808a189eba8836efb7f28d0" gracePeriod=30 Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.671794 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.671830 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.724872 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95edebcc-4231-4f68-be9d-7e04a0d15e16-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.724932 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwx5g\" (UniqueName: \"kubernetes.io/projected/95edebcc-4231-4f68-be9d-7e04a0d15e16-kube-api-access-wwx5g\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.725088 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95edebcc-4231-4f68-be9d-7e04a0d15e16-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.729080 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95edebcc-4231-4f68-be9d-7e04a0d15e16-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.729333 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95edebcc-4231-4f68-be9d-7e04a0d15e16-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.744315 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwx5g\" (UniqueName: \"kubernetes.io/projected/95edebcc-4231-4f68-be9d-7e04a0d15e16-kube-api-access-wwx5g\") pod \"nova-cell1-conductor-0\" (UID: \"95edebcc-4231-4f68-be9d-7e04a0d15e16\") " pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.849111 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:56 crc kubenswrapper[4679]: I1202 10:40:56.920466 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d" path="/var/lib/kubelet/pods/4fd5e0ee-bc7f-48de-9d1c-f15f8d98e27d/volumes" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.398580 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.428404 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"95edebcc-4231-4f68-be9d-7e04a0d15e16","Type":"ContainerStarted","Data":"2194521ecd6a6289b77e093cca0963da4d13c81a9d4b456f18110da5dd0f8e16"} Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.432737 4679 generic.go:334] "Generic (PLEG): container finished" podID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerID="09a5370e9d2bb3ccead6f3ab1604562f5c2d5842e808a189eba8836efb7f28d0" exitCode=0 Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.432766 4679 generic.go:334] "Generic (PLEG): container finished" podID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerID="51fbfe9ca51bb63f94d28d8e629804c8b943dbeee498ec2291f9894d9975073c" exitCode=143 Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.432800 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe","Type":"ContainerDied","Data":"09a5370e9d2bb3ccead6f3ab1604562f5c2d5842e808a189eba8836efb7f28d0"} Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.432823 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe","Type":"ContainerDied","Data":"51fbfe9ca51bb63f94d28d8e629804c8b943dbeee498ec2291f9894d9975073c"} Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.432835 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe","Type":"ContainerDied","Data":"b6ab9cca5b6222eceddc98fb5bf46663fba3e83696db056993f7de75fdcadbd9"} Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.432844 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6ab9cca5b6222eceddc98fb5bf46663fba3e83696db056993f7de75fdcadbd9" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.439176 4679 generic.go:334] "Generic (PLEG): container finished" podID="d17126b2-501f-4d23-96fa-91d694341f9b" containerID="96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd" exitCode=143 Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.439414 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="dee36d9b-8726-49b5-b9d2-bb695c3442c3" containerName="nova-scheduler-scheduler" containerID="cri-o://caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0" gracePeriod=30 Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.440515 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d17126b2-501f-4d23-96fa-91d694341f9b","Type":"ContainerDied","Data":"96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd"} Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.585965 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.748894 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-combined-ca-bundle\") pod \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.748974 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-config-data\") pod \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.749037 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-logs\") pod \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.749144 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-nova-metadata-tls-certs\") pod \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.749351 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf5kp\" (UniqueName: \"kubernetes.io/projected/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-kube-api-access-vf5kp\") pod \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\" (UID: \"13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe\") " Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.749529 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-logs" (OuterVolumeSpecName: "logs") pod "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" (UID: "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.750378 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.754554 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-kube-api-access-vf5kp" (OuterVolumeSpecName: "kube-api-access-vf5kp") pod "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" (UID: "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe"). InnerVolumeSpecName "kube-api-access-vf5kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.777177 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" (UID: "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.782151 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-config-data" (OuterVolumeSpecName: "config-data") pod "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" (UID: "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.815255 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" (UID: "13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.851562 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf5kp\" (UniqueName: \"kubernetes.io/projected/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-kube-api-access-vf5kp\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.851592 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.851602 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:57 crc kubenswrapper[4679]: I1202 10:40:57.851612 4679 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.449055 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.450210 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"95edebcc-4231-4f68-be9d-7e04a0d15e16","Type":"ContainerStarted","Data":"e8462fac16655e0984b266bfa787cb7211dc2b02ad444630123fc51dc1d1a46f"} Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.450614 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.475468 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.475450603 podStartE2EDuration="2.475450603s" podCreationTimestamp="2025-12-02 10:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:40:58.471494959 +0000 UTC m=+1311.801633819" watchObservedRunningTime="2025-12-02 10:40:58.475450603 +0000 UTC m=+1311.805589463" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.493296 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.504322 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.514767 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:58 crc kubenswrapper[4679]: E1202 10:40:58.523418 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-metadata" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.523559 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-metadata" Dec 02 10:40:58 crc kubenswrapper[4679]: E1202 10:40:58.523696 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-log" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.523769 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-log" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.524078 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-metadata" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.524201 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" containerName="nova-metadata-log" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.527213 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.529731 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.529920 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.560146 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.675295 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-config-data\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.675406 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-logs\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.675436 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd7q2\" (UniqueName: \"kubernetes.io/projected/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-kube-api-access-gd7q2\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.675463 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.675830 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.777747 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.777893 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-config-data\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.777938 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-logs\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.777958 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd7q2\" (UniqueName: \"kubernetes.io/projected/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-kube-api-access-gd7q2\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.777984 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.780221 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-logs\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.784190 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.785141 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-config-data\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.788661 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.801937 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd7q2\" (UniqueName: \"kubernetes.io/projected/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-kube-api-access-gd7q2\") pod \"nova-metadata-0\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.895651 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:40:58 crc kubenswrapper[4679]: I1202 10:40:58.932409 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe" path="/var/lib/kubelet/pods/13b59b5e-6c16-48e0-9e32-a4f4c9fd97fe/volumes" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.225670 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.288646 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-config-data\") pod \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.288855 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-combined-ca-bundle\") pod \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.288962 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnbb4\" (UniqueName: \"kubernetes.io/projected/dee36d9b-8726-49b5-b9d2-bb695c3442c3-kube-api-access-bnbb4\") pod \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\" (UID: \"dee36d9b-8726-49b5-b9d2-bb695c3442c3\") " Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.311898 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dee36d9b-8726-49b5-b9d2-bb695c3442c3-kube-api-access-bnbb4" (OuterVolumeSpecName: "kube-api-access-bnbb4") pod "dee36d9b-8726-49b5-b9d2-bb695c3442c3" (UID: "dee36d9b-8726-49b5-b9d2-bb695c3442c3"). InnerVolumeSpecName "kube-api-access-bnbb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.322926 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-config-data" (OuterVolumeSpecName: "config-data") pod "dee36d9b-8726-49b5-b9d2-bb695c3442c3" (UID: "dee36d9b-8726-49b5-b9d2-bb695c3442c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.324970 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dee36d9b-8726-49b5-b9d2-bb695c3442c3" (UID: "dee36d9b-8726-49b5-b9d2-bb695c3442c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.391038 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.391327 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnbb4\" (UniqueName: \"kubernetes.io/projected/dee36d9b-8726-49b5-b9d2-bb695c3442c3-kube-api-access-bnbb4\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.391339 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee36d9b-8726-49b5-b9d2-bb695c3442c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.416787 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.458833 4679 generic.go:334] "Generic (PLEG): container finished" podID="dee36d9b-8726-49b5-b9d2-bb695c3442c3" containerID="caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0" exitCode=0 Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.458962 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.460375 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dee36d9b-8726-49b5-b9d2-bb695c3442c3","Type":"ContainerDied","Data":"caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0"} Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.460441 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dee36d9b-8726-49b5-b9d2-bb695c3442c3","Type":"ContainerDied","Data":"4b0974d3052b6ce3c8e36e4cf1d63d5e954f86e8065b308a6e308a25b0c1f9da"} Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.460463 4679 scope.go:117] "RemoveContainer" containerID="caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.464399 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4bc7924b-f6dd-4fb9-bf1c-64af7274a161","Type":"ContainerStarted","Data":"44018de466426ced3419c9378ebb562670aa7438620f753bc45ac75c15e5b82a"} Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.490843 4679 scope.go:117] "RemoveContainer" containerID="caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0" Dec 02 10:40:59 crc kubenswrapper[4679]: E1202 10:40:59.491439 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0\": container with ID starting with caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0 not found: ID does not exist" containerID="caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.491469 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0"} err="failed to get container status \"caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0\": rpc error: code = NotFound desc = could not find container \"caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0\": container with ID starting with caefb66f999765a65c091d7cb5c3606c7f54b2002ad087346356aac2022e47a0 not found: ID does not exist" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.501142 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.516979 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.525767 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:59 crc kubenswrapper[4679]: E1202 10:40:59.526388 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dee36d9b-8726-49b5-b9d2-bb695c3442c3" containerName="nova-scheduler-scheduler" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.526402 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="dee36d9b-8726-49b5-b9d2-bb695c3442c3" containerName="nova-scheduler-scheduler" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.526587 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="dee36d9b-8726-49b5-b9d2-bb695c3442c3" containerName="nova-scheduler-scheduler" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.527145 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.529788 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.551994 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.596480 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f8cg\" (UniqueName: \"kubernetes.io/projected/4c7bbbe3-232a-41b2-9378-848755f2d1e6-kube-api-access-4f8cg\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.596534 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.596763 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-config-data\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.698604 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-config-data\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.698677 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f8cg\" (UniqueName: \"kubernetes.io/projected/4c7bbbe3-232a-41b2-9378-848755f2d1e6-kube-api-access-4f8cg\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.698704 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.710891 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-config-data\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.713123 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.720543 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f8cg\" (UniqueName: \"kubernetes.io/projected/4c7bbbe3-232a-41b2-9378-848755f2d1e6-kube-api-access-4f8cg\") pod \"nova-scheduler-0\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " pod="openstack/nova-scheduler-0" Dec 02 10:40:59 crc kubenswrapper[4679]: I1202 10:40:59.862681 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.409938 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:00 crc kubenswrapper[4679]: W1202 10:41:00.420694 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c7bbbe3_232a_41b2_9378_848755f2d1e6.slice/crio-a8674757fc33c9eb00f02c99c4539b155d88db691351c9636a69fb4bac28fc35 WatchSource:0}: Error finding container a8674757fc33c9eb00f02c99c4539b155d88db691351c9636a69fb4bac28fc35: Status 404 returned error can't find the container with id a8674757fc33c9eb00f02c99c4539b155d88db691351c9636a69fb4bac28fc35 Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.475592 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4c7bbbe3-232a-41b2-9378-848755f2d1e6","Type":"ContainerStarted","Data":"a8674757fc33c9eb00f02c99c4539b155d88db691351c9636a69fb4bac28fc35"} Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.479676 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4bc7924b-f6dd-4fb9-bf1c-64af7274a161","Type":"ContainerStarted","Data":"bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf"} Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.479720 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4bc7924b-f6dd-4fb9-bf1c-64af7274a161","Type":"ContainerStarted","Data":"ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf"} Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.485851 4679 generic.go:334] "Generic (PLEG): container finished" podID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerID="e8ec769056d0a018244b7c3776f0ce7019c0e08a27bb5bb230aed17faccebd1a" exitCode=0 Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.485890 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerDied","Data":"e8ec769056d0a018244b7c3776f0ce7019c0e08a27bb5bb230aed17faccebd1a"} Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.505840 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.50581704 podStartE2EDuration="2.50581704s" podCreationTimestamp="2025-12-02 10:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:00.498983943 +0000 UTC m=+1313.829122803" watchObservedRunningTime="2025-12-02 10:41:00.50581704 +0000 UTC m=+1313.835955900" Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.829944 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 02 10:41:00 crc kubenswrapper[4679]: I1202 10:41:00.922685 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dee36d9b-8726-49b5-b9d2-bb695c3442c3" path="/var/lib/kubelet/pods/dee36d9b-8726-49b5-b9d2-bb695c3442c3/volumes" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.441215 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.445842 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.497175 4679 generic.go:334] "Generic (PLEG): container finished" podID="d17126b2-501f-4d23-96fa-91d694341f9b" containerID="b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df" exitCode=0 Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.497226 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.497219 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d17126b2-501f-4d23-96fa-91d694341f9b","Type":"ContainerDied","Data":"b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df"} Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.497687 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d17126b2-501f-4d23-96fa-91d694341f9b","Type":"ContainerDied","Data":"1140491dd2e3a1df97f0c05f5bf4eea61305f43fd6faf541b4bb0a62592b5070"} Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.497731 4679 scope.go:117] "RemoveContainer" containerID="b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.504122 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ce2ce724-3129-47f2-a47a-c73e9f5576cd","Type":"ContainerDied","Data":"0739173764fa90fdf3b0627ff8d9eea8789f8c5dc1105fecc681fb941812e60f"} Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.504154 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.514627 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4c7bbbe3-232a-41b2-9378-848755f2d1e6","Type":"ContainerStarted","Data":"5e9e6d52b8d0c88e455191d96c76b4befb97c7f8b7ffb0ffee8623f646d8a7d5"} Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.533935 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-config-data\") pod \"d17126b2-501f-4d23-96fa-91d694341f9b\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534003 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-config-data\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534072 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-combined-ca-bundle\") pod \"d17126b2-501f-4d23-96fa-91d694341f9b\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534116 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-scripts\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534182 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17126b2-501f-4d23-96fa-91d694341f9b-logs\") pod \"d17126b2-501f-4d23-96fa-91d694341f9b\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534261 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-log-httpd\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534326 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7knr\" (UniqueName: \"kubernetes.io/projected/d17126b2-501f-4d23-96fa-91d694341f9b-kube-api-access-w7knr\") pod \"d17126b2-501f-4d23-96fa-91d694341f9b\" (UID: \"d17126b2-501f-4d23-96fa-91d694341f9b\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534366 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8stdz\" (UniqueName: \"kubernetes.io/projected/ce2ce724-3129-47f2-a47a-c73e9f5576cd-kube-api-access-8stdz\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534421 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-run-httpd\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534491 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-sg-core-conf-yaml\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.534542 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-combined-ca-bundle\") pod \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\" (UID: \"ce2ce724-3129-47f2-a47a-c73e9f5576cd\") " Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.537686 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17126b2-501f-4d23-96fa-91d694341f9b-logs" (OuterVolumeSpecName: "logs") pod "d17126b2-501f-4d23-96fa-91d694341f9b" (UID: "d17126b2-501f-4d23-96fa-91d694341f9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.538552 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.538630 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.539020 4679 scope.go:117] "RemoveContainer" containerID="96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.577681 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-scripts" (OuterVolumeSpecName: "scripts") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.593629 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce2ce724-3129-47f2-a47a-c73e9f5576cd-kube-api-access-8stdz" (OuterVolumeSpecName: "kube-api-access-8stdz") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "kube-api-access-8stdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.593812 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17126b2-501f-4d23-96fa-91d694341f9b-kube-api-access-w7knr" (OuterVolumeSpecName: "kube-api-access-w7knr") pod "d17126b2-501f-4d23-96fa-91d694341f9b" (UID: "d17126b2-501f-4d23-96fa-91d694341f9b"). InnerVolumeSpecName "kube-api-access-w7knr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.613715 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d17126b2-501f-4d23-96fa-91d694341f9b" (UID: "d17126b2-501f-4d23-96fa-91d694341f9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.615675 4679 scope.go:117] "RemoveContainer" containerID="b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.616766 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df\": container with ID starting with b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df not found: ID does not exist" containerID="b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.616833 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df"} err="failed to get container status \"b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df\": rpc error: code = NotFound desc = could not find container \"b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df\": container with ID starting with b36c6ac542ff0ba58221721792f0d0c9a0bb3d41185b779b8fb37f726461f2df not found: ID does not exist" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.616869 4679 scope.go:117] "RemoveContainer" containerID="96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.617276 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd\": container with ID starting with 96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd not found: ID does not exist" containerID="96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.617345 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd"} err="failed to get container status \"96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd\": rpc error: code = NotFound desc = could not find container \"96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd\": container with ID starting with 96b6e07c9acce840beffaa9c2faf10ce77f489feb255922833c46cc470942efd not found: ID does not exist" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.617373 4679 scope.go:117] "RemoveContainer" containerID="c2dc0e5e1cbf278d9e3bb348987e252d165a74f68d43b1422d0c5ae6d93900e8" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.626092 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.638392 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-config-data" (OuterVolumeSpecName: "config-data") pod "d17126b2-501f-4d23-96fa-91d694341f9b" (UID: "d17126b2-501f-4d23-96fa-91d694341f9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.639994 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640111 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640195 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17126b2-501f-4d23-96fa-91d694341f9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640282 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640425 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17126b2-501f-4d23-96fa-91d694341f9b-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640505 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640589 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7knr\" (UniqueName: \"kubernetes.io/projected/d17126b2-501f-4d23-96fa-91d694341f9b-kube-api-access-w7knr\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640664 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8stdz\" (UniqueName: \"kubernetes.io/projected/ce2ce724-3129-47f2-a47a-c73e9f5576cd-kube-api-access-8stdz\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.640740 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ce2ce724-3129-47f2-a47a-c73e9f5576cd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.654983 4679 scope.go:117] "RemoveContainer" containerID="e84ae32bcfe5a9158e3f189a9abaf5c62f03f679d72ef1657aa70057953f2f68" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.679620 4679 scope.go:117] "RemoveContainer" containerID="e8ec769056d0a018244b7c3776f0ce7019c0e08a27bb5bb230aed17faccebd1a" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.684382 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.684467 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-config-data" (OuterVolumeSpecName: "config-data") pod "ce2ce724-3129-47f2-a47a-c73e9f5576cd" (UID: "ce2ce724-3129-47f2-a47a-c73e9f5576cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.697101 4679 scope.go:117] "RemoveContainer" containerID="1d8ba8e5aa030cca61f76f05247e5c15824e6017e04f7cd9a5540679004eb17d" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.742424 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.742455 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2ce724-3129-47f2-a47a-c73e9f5576cd-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.825124 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.825106914 podStartE2EDuration="2.825106914s" podCreationTimestamp="2025-12-02 10:40:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:01.544176888 +0000 UTC m=+1314.874315748" watchObservedRunningTime="2025-12-02 10:41:01.825106914 +0000 UTC m=+1315.155245764" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.843217 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.868642 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.895450 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.904066 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.911801 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.912165 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="sg-core" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912182 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="sg-core" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.912197 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-api" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912203 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-api" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.912240 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-central-agent" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912246 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-central-agent" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.912257 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="proxy-httpd" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912263 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="proxy-httpd" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.912273 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-log" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912278 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-log" Dec 02 10:41:01 crc kubenswrapper[4679]: E1202 10:41:01.912289 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-notification-agent" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912296 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-notification-agent" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912544 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-central-agent" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912560 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="ceilometer-notification-agent" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912578 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="proxy-httpd" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912587 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-log" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912598 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" containerName="sg-core" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.912611 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" containerName="nova-api-api" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.913672 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.916409 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.919547 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.921913 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.924785 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.924972 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.925689 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.927900 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:01 crc kubenswrapper[4679]: I1202 10:41:01.935237 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.046771 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n6h5\" (UniqueName: \"kubernetes.io/projected/ff686c4d-3ada-476a-8c5b-85774b8be5bc-kube-api-access-8n6h5\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.046818 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-run-httpd\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.046850 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff686c4d-3ada-476a-8c5b-85774b8be5bc-logs\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.046870 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047100 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-log-httpd\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047243 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047283 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bnft\" (UniqueName: \"kubernetes.io/projected/e31ec677-fadd-4f7d-944a-ed48fdc5b094-kube-api-access-8bnft\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047323 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-scripts\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047445 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047509 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-config-data\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047748 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.047846 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-config-data\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150002 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150044 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-scripts\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150071 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bnft\" (UniqueName: \"kubernetes.io/projected/e31ec677-fadd-4f7d-944a-ed48fdc5b094-kube-api-access-8bnft\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150120 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150157 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-config-data\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150219 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150241 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-config-data\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150275 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n6h5\" (UniqueName: \"kubernetes.io/projected/ff686c4d-3ada-476a-8c5b-85774b8be5bc-kube-api-access-8n6h5\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150294 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-run-httpd\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150347 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff686c4d-3ada-476a-8c5b-85774b8be5bc-logs\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150368 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150414 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-log-httpd\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.150830 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-log-httpd\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.151038 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff686c4d-3ada-476a-8c5b-85774b8be5bc-logs\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.151038 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-run-httpd\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.154482 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.154853 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.154919 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-scripts\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.155041 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.155210 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-config-data\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.155456 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.156485 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-config-data\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.170921 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n6h5\" (UniqueName: \"kubernetes.io/projected/ff686c4d-3ada-476a-8c5b-85774b8be5bc-kube-api-access-8n6h5\") pod \"nova-api-0\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.171157 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bnft\" (UniqueName: \"kubernetes.io/projected/e31ec677-fadd-4f7d-944a-ed48fdc5b094-kube-api-access-8bnft\") pod \"ceilometer-0\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.247580 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.263427 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.779154 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.788876 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:02 crc kubenswrapper[4679]: W1202 10:41:02.805415 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode31ec677_fadd_4f7d_944a_ed48fdc5b094.slice/crio-1e16739f94af62aec4c3d69a99dd404f22ab748d99caf0035fbdcfb4493421fc WatchSource:0}: Error finding container 1e16739f94af62aec4c3d69a99dd404f22ab748d99caf0035fbdcfb4493421fc: Status 404 returned error can't find the container with id 1e16739f94af62aec4c3d69a99dd404f22ab748d99caf0035fbdcfb4493421fc Dec 02 10:41:02 crc kubenswrapper[4679]: W1202 10:41:02.810196 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff686c4d_3ada_476a_8c5b_85774b8be5bc.slice/crio-a52241d780bab9e7fafee846e71810db3ec89f849bb03184e21028c7043e8dc3 WatchSource:0}: Error finding container a52241d780bab9e7fafee846e71810db3ec89f849bb03184e21028c7043e8dc3: Status 404 returned error can't find the container with id a52241d780bab9e7fafee846e71810db3ec89f849bb03184e21028c7043e8dc3 Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.929298 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce2ce724-3129-47f2-a47a-c73e9f5576cd" path="/var/lib/kubelet/pods/ce2ce724-3129-47f2-a47a-c73e9f5576cd/volumes" Dec 02 10:41:02 crc kubenswrapper[4679]: I1202 10:41:02.930647 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17126b2-501f-4d23-96fa-91d694341f9b" path="/var/lib/kubelet/pods/d17126b2-501f-4d23-96fa-91d694341f9b/volumes" Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.536163 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff686c4d-3ada-476a-8c5b-85774b8be5bc","Type":"ContainerStarted","Data":"0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966"} Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.536213 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff686c4d-3ada-476a-8c5b-85774b8be5bc","Type":"ContainerStarted","Data":"5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e"} Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.536224 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff686c4d-3ada-476a-8c5b-85774b8be5bc","Type":"ContainerStarted","Data":"a52241d780bab9e7fafee846e71810db3ec89f849bb03184e21028c7043e8dc3"} Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.537662 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerStarted","Data":"1e16739f94af62aec4c3d69a99dd404f22ab748d99caf0035fbdcfb4493421fc"} Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.560397 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.560376318 podStartE2EDuration="2.560376318s" podCreationTimestamp="2025-12-02 10:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:03.553608543 +0000 UTC m=+1316.883747403" watchObservedRunningTime="2025-12-02 10:41:03.560376318 +0000 UTC m=+1316.890515178" Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.896575 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:41:03 crc kubenswrapper[4679]: I1202 10:41:03.896644 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:41:04 crc kubenswrapper[4679]: I1202 10:41:04.547148 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerStarted","Data":"ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee"} Dec 02 10:41:04 crc kubenswrapper[4679]: I1202 10:41:04.863815 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 10:41:05 crc kubenswrapper[4679]: I1202 10:41:05.557949 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerStarted","Data":"c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972"} Dec 02 10:41:06 crc kubenswrapper[4679]: I1202 10:41:06.569434 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerStarted","Data":"b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f"} Dec 02 10:41:06 crc kubenswrapper[4679]: I1202 10:41:06.878835 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 02 10:41:08 crc kubenswrapper[4679]: I1202 10:41:08.611001 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerStarted","Data":"307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795"} Dec 02 10:41:08 crc kubenswrapper[4679]: I1202 10:41:08.611689 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:41:08 crc kubenswrapper[4679]: I1202 10:41:08.648994 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.716488595 podStartE2EDuration="7.648968968s" podCreationTimestamp="2025-12-02 10:41:01 +0000 UTC" firstStartedPulling="2025-12-02 10:41:02.811108969 +0000 UTC m=+1316.141247829" lastFinishedPulling="2025-12-02 10:41:07.743589342 +0000 UTC m=+1321.073728202" observedRunningTime="2025-12-02 10:41:08.633818373 +0000 UTC m=+1321.963957243" watchObservedRunningTime="2025-12-02 10:41:08.648968968 +0000 UTC m=+1321.979107828" Dec 02 10:41:08 crc kubenswrapper[4679]: I1202 10:41:08.896767 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 10:41:08 crc kubenswrapper[4679]: I1202 10:41:08.896818 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 10:41:09 crc kubenswrapper[4679]: I1202 10:41:09.863564 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 10:41:09 crc kubenswrapper[4679]: I1202 10:41:09.896145 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 10:41:09 crc kubenswrapper[4679]: I1202 10:41:09.910476 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:09 crc kubenswrapper[4679]: I1202 10:41:09.910498 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:10 crc kubenswrapper[4679]: I1202 10:41:10.659983 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 10:41:12 crc kubenswrapper[4679]: I1202 10:41:12.248031 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:41:12 crc kubenswrapper[4679]: I1202 10:41:12.248399 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:41:13 crc kubenswrapper[4679]: I1202 10:41:13.330589 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:13 crc kubenswrapper[4679]: I1202 10:41:13.330591 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:18 crc kubenswrapper[4679]: I1202 10:41:18.903422 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 10:41:18 crc kubenswrapper[4679]: I1202 10:41:18.906608 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 10:41:18 crc kubenswrapper[4679]: I1202 10:41:18.922941 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 10:41:19 crc kubenswrapper[4679]: I1202 10:41:19.708547 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.674952 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.712406 4679 generic.go:334] "Generic (PLEG): container finished" podID="17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" containerID="a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1" exitCode=137 Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.712465 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d","Type":"ContainerDied","Data":"a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1"} Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.712480 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.712498 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d","Type":"ContainerDied","Data":"f8447faa31284d856fee03e0a1dcccf22679fd4add6ac2c479e0e36c6d5d1a85"} Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.712528 4679 scope.go:117] "RemoveContainer" containerID="a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.736414 4679 scope.go:117] "RemoveContainer" containerID="a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1" Dec 02 10:41:20 crc kubenswrapper[4679]: E1202 10:41:20.736906 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1\": container with ID starting with a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1 not found: ID does not exist" containerID="a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.736943 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1"} err="failed to get container status \"a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1\": rpc error: code = NotFound desc = could not find container \"a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1\": container with ID starting with a517939da2bd248360334c938819ad7a4a7b445072f7de12c3cb2b1aae926dd1 not found: ID does not exist" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.806105 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-config-data\") pod \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.806479 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppfrb\" (UniqueName: \"kubernetes.io/projected/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-kube-api-access-ppfrb\") pod \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.806533 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-combined-ca-bundle\") pod \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\" (UID: \"17eeecb2-5cdc-4eaa-aac6-a0939f8e003d\") " Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.811530 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-kube-api-access-ppfrb" (OuterVolumeSpecName: "kube-api-access-ppfrb") pod "17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" (UID: "17eeecb2-5cdc-4eaa-aac6-a0939f8e003d"). InnerVolumeSpecName "kube-api-access-ppfrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.836184 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-config-data" (OuterVolumeSpecName: "config-data") pod "17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" (UID: "17eeecb2-5cdc-4eaa-aac6-a0939f8e003d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.837526 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" (UID: "17eeecb2-5cdc-4eaa-aac6-a0939f8e003d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.910888 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.910937 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppfrb\" (UniqueName: \"kubernetes.io/projected/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-kube-api-access-ppfrb\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:20 crc kubenswrapper[4679]: I1202 10:41:20.910954 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.038077 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.053288 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.061627 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:41:21 crc kubenswrapper[4679]: E1202 10:41:21.062033 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.062052 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.062256 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.062853 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.068740 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.070387 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.070707 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.072640 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.218724 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58cgj\" (UniqueName: \"kubernetes.io/projected/30a95365-228b-4db8-bcae-4957f4a4ee87-kube-api-access-58cgj\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.219102 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.219182 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.219212 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.219412 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.321106 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.321236 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.321267 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.321299 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.321358 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58cgj\" (UniqueName: \"kubernetes.io/projected/30a95365-228b-4db8-bcae-4957f4a4ee87-kube-api-access-58cgj\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.325336 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.325681 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.326663 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.330926 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a95365-228b-4db8-bcae-4957f4a4ee87-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.339576 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58cgj\" (UniqueName: \"kubernetes.io/projected/30a95365-228b-4db8-bcae-4957f4a4ee87-kube-api-access-58cgj\") pod \"nova-cell1-novncproxy-0\" (UID: \"30a95365-228b-4db8-bcae-4957f4a4ee87\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.378602 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:21 crc kubenswrapper[4679]: I1202 10:41:21.837542 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.253721 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.255394 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.255608 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.258107 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.738334 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30a95365-228b-4db8-bcae-4957f4a4ee87","Type":"ContainerStarted","Data":"06caff582b79d03d3b552e8fad449955af701094ba21395b47b96841a68852ed"} Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.738384 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30a95365-228b-4db8-bcae-4957f4a4ee87","Type":"ContainerStarted","Data":"86d473ac1a09090aae81dce40d6a7a94f8dcbe50f7f3cc94e3f7c4fede6ab6f1"} Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.738749 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.743329 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.758009 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.7579918060000002 podStartE2EDuration="1.757991806s" podCreationTimestamp="2025-12-02 10:41:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:22.753697372 +0000 UTC m=+1336.083836242" watchObservedRunningTime="2025-12-02 10:41:22.757991806 +0000 UTC m=+1336.088130666" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.953206 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17eeecb2-5cdc-4eaa-aac6-a0939f8e003d" path="/var/lib/kubelet/pods/17eeecb2-5cdc-4eaa-aac6-a0939f8e003d/volumes" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.954179 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-kj97s"] Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.956317 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:22 crc kubenswrapper[4679]: I1202 10:41:22.977844 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-kj97s"] Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.056078 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.056140 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.056197 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.056215 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-config\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.056236 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.056278 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4dj8\" (UniqueName: \"kubernetes.io/projected/03f46a70-31b1-49ee-ab7a-ebd520592409-kube-api-access-l4dj8\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.157548 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.157615 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.157674 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.157694 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-config\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.157718 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.157765 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4dj8\" (UniqueName: \"kubernetes.io/projected/03f46a70-31b1-49ee-ab7a-ebd520592409-kube-api-access-l4dj8\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.159356 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.159437 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.159562 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-config\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.159781 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.160187 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.179783 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4dj8\" (UniqueName: \"kubernetes.io/projected/03f46a70-31b1-49ee-ab7a-ebd520592409-kube-api-access-l4dj8\") pod \"dnsmasq-dns-59cf4bdb65-kj97s\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.289065 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:23 crc kubenswrapper[4679]: I1202 10:41:23.820264 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-kj97s"] Dec 02 10:41:24 crc kubenswrapper[4679]: I1202 10:41:24.757883 4679 generic.go:334] "Generic (PLEG): container finished" podID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerID="a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5" exitCode=0 Dec 02 10:41:24 crc kubenswrapper[4679]: I1202 10:41:24.758053 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" event={"ID":"03f46a70-31b1-49ee-ab7a-ebd520592409","Type":"ContainerDied","Data":"a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5"} Dec 02 10:41:24 crc kubenswrapper[4679]: I1202 10:41:24.758879 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" event={"ID":"03f46a70-31b1-49ee-ab7a-ebd520592409","Type":"ContainerStarted","Data":"eeb85bfba02139b8245af19e5a8a7532cd217ecd23fb63acf596f22c02fa6bf5"} Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.128638 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.129324 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-central-agent" containerID="cri-o://ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee" gracePeriod=30 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.129357 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="sg-core" containerID="cri-o://b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f" gracePeriod=30 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.129409 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-notification-agent" containerID="cri-o://c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972" gracePeriod=30 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.129371 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="proxy-httpd" containerID="cri-o://307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795" gracePeriod=30 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.137437 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": read tcp 10.217.0.2:42858->10.217.0.194:3000: read: connection reset by peer" Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.531842 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.773534 4679 generic.go:334] "Generic (PLEG): container finished" podID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerID="307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795" exitCode=0 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.773834 4679 generic.go:334] "Generic (PLEG): container finished" podID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerID="b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f" exitCode=2 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.773849 4679 generic.go:334] "Generic (PLEG): container finished" podID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerID="ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee" exitCode=0 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.773910 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerDied","Data":"307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795"} Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.773940 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerDied","Data":"b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f"} Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.773954 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerDied","Data":"ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee"} Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.776135 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-log" containerID="cri-o://5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e" gracePeriod=30 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.776517 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-api" containerID="cri-o://0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966" gracePeriod=30 Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.776830 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" event={"ID":"03f46a70-31b1-49ee-ab7a-ebd520592409","Type":"ContainerStarted","Data":"b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a"} Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.776974 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:25 crc kubenswrapper[4679]: I1202 10:41:25.815939 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" podStartSLOduration=3.815298424 podStartE2EDuration="3.815298424s" podCreationTimestamp="2025-12-02 10:41:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:25.805466812 +0000 UTC m=+1339.135605682" watchObservedRunningTime="2025-12-02 10:41:25.815298424 +0000 UTC m=+1339.145437284" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.361514 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.379188 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.421198 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-scripts\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.421682 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-config-data\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.421813 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-combined-ca-bundle\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.421841 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-ceilometer-tls-certs\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.422564 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-sg-core-conf-yaml\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.422671 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-run-httpd\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.422695 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bnft\" (UniqueName: \"kubernetes.io/projected/e31ec677-fadd-4f7d-944a-ed48fdc5b094-kube-api-access-8bnft\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.422759 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-log-httpd\") pod \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\" (UID: \"e31ec677-fadd-4f7d-944a-ed48fdc5b094\") " Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.428167 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.428193 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.525107 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.525142 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e31ec677-fadd-4f7d-944a-ed48fdc5b094-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.540668 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-scripts" (OuterVolumeSpecName: "scripts") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.541047 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31ec677-fadd-4f7d-944a-ed48fdc5b094-kube-api-access-8bnft" (OuterVolumeSpecName: "kube-api-access-8bnft") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "kube-api-access-8bnft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.553409 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.605413 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.606434 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.623427 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-config-data" (OuterVolumeSpecName: "config-data") pod "e31ec677-fadd-4f7d-944a-ed48fdc5b094" (UID: "e31ec677-fadd-4f7d-944a-ed48fdc5b094"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.627219 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bnft\" (UniqueName: \"kubernetes.io/projected/e31ec677-fadd-4f7d-944a-ed48fdc5b094-kube-api-access-8bnft\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.627275 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.627290 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.627320 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.627331 4679 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.627342 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e31ec677-fadd-4f7d-944a-ed48fdc5b094-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.785239 4679 generic.go:334] "Generic (PLEG): container finished" podID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerID="5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e" exitCode=143 Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.785334 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff686c4d-3ada-476a-8c5b-85774b8be5bc","Type":"ContainerDied","Data":"5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e"} Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.787791 4679 generic.go:334] "Generic (PLEG): container finished" podID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerID="c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972" exitCode=0 Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.788791 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.797378 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerDied","Data":"c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972"} Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.797457 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e31ec677-fadd-4f7d-944a-ed48fdc5b094","Type":"ContainerDied","Data":"1e16739f94af62aec4c3d69a99dd404f22ab748d99caf0035fbdcfb4493421fc"} Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.797477 4679 scope.go:117] "RemoveContainer" containerID="307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.825286 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.849537 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.855280 4679 scope.go:117] "RemoveContainer" containerID="b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.870820 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.871294 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-notification-agent" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871333 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-notification-agent" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.871353 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="proxy-httpd" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871359 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="proxy-httpd" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.871382 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="sg-core" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871405 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="sg-core" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.871424 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-central-agent" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871432 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-central-agent" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871677 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-central-agent" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871691 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="ceilometer-notification-agent" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871702 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="sg-core" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.871735 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" containerName="proxy-httpd" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.874389 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.878934 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.879145 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.879575 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.890611 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.893853 4679 scope.go:117] "RemoveContainer" containerID="c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.921669 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31ec677-fadd-4f7d-944a-ed48fdc5b094" path="/var/lib/kubelet/pods/e31ec677-fadd-4f7d-944a-ed48fdc5b094/volumes" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.924770 4679 scope.go:117] "RemoveContainer" containerID="ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.931886 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.931939 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-config-data\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.932747 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-log-httpd\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.932830 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.932971 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-run-httpd\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.933208 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-scripts\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.933369 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.933446 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d284z\" (UniqueName: \"kubernetes.io/projected/24c4ed59-db75-4d18-9889-bfea9ca275b7-kube-api-access-d284z\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.941032 4679 scope.go:117] "RemoveContainer" containerID="307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.941485 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795\": container with ID starting with 307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795 not found: ID does not exist" containerID="307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.941532 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795"} err="failed to get container status \"307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795\": rpc error: code = NotFound desc = could not find container \"307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795\": container with ID starting with 307328e87158e7d41b3687b33933e52f6d086625ab13208721cd3b851b089795 not found: ID does not exist" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.941564 4679 scope.go:117] "RemoveContainer" containerID="b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.941843 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f\": container with ID starting with b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f not found: ID does not exist" containerID="b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.941873 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f"} err="failed to get container status \"b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f\": rpc error: code = NotFound desc = could not find container \"b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f\": container with ID starting with b4fbf212e87be1a02a6dff2c432f44b324da88a682babb346899f4d105d8c07f not found: ID does not exist" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.941892 4679 scope.go:117] "RemoveContainer" containerID="c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.942089 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972\": container with ID starting with c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972 not found: ID does not exist" containerID="c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.942116 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972"} err="failed to get container status \"c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972\": rpc error: code = NotFound desc = could not find container \"c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972\": container with ID starting with c4b5fe60af698f4583d85d7e8a00ea53a9fcef8518ae8b4537587a94db3b3972 not found: ID does not exist" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.942133 4679 scope.go:117] "RemoveContainer" containerID="ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee" Dec 02 10:41:26 crc kubenswrapper[4679]: E1202 10:41:26.942326 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee\": container with ID starting with ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee not found: ID does not exist" containerID="ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee" Dec 02 10:41:26 crc kubenswrapper[4679]: I1202 10:41:26.942353 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee"} err="failed to get container status \"ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee\": rpc error: code = NotFound desc = could not find container \"ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee\": container with ID starting with ec12e07de6be991352c62060414e6a976c5bb0dea10487f9037d14a2395faeee not found: ID does not exist" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.035965 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-log-httpd\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036009 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036045 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-run-httpd\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036127 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-scripts\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036155 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036174 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d284z\" (UniqueName: \"kubernetes.io/projected/24c4ed59-db75-4d18-9889-bfea9ca275b7-kube-api-access-d284z\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036245 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036279 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-config-data\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.036382 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-log-httpd\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.037271 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-run-httpd\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.041079 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-scripts\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.041751 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.042372 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-config-data\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.043401 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.050902 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.054845 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d284z\" (UniqueName: \"kubernetes.io/projected/24c4ed59-db75-4d18-9889-bfea9ca275b7-kube-api-access-d284z\") pod \"ceilometer-0\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.078686 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.080373 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.587360 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.594406 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:41:27 crc kubenswrapper[4679]: I1202 10:41:27.800915 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerStarted","Data":"1d126fd2fef35eb431c032e6e1982dd6c58ddedbdad79efcb39adbb174750927"} Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.425876 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.589048 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff686c4d-3ada-476a-8c5b-85774b8be5bc-logs\") pod \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.589413 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-combined-ca-bundle\") pod \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.589496 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n6h5\" (UniqueName: \"kubernetes.io/projected/ff686c4d-3ada-476a-8c5b-85774b8be5bc-kube-api-access-8n6h5\") pod \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.589528 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-config-data\") pod \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\" (UID: \"ff686c4d-3ada-476a-8c5b-85774b8be5bc\") " Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.589785 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff686c4d-3ada-476a-8c5b-85774b8be5bc-logs" (OuterVolumeSpecName: "logs") pod "ff686c4d-3ada-476a-8c5b-85774b8be5bc" (UID: "ff686c4d-3ada-476a-8c5b-85774b8be5bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.589997 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff686c4d-3ada-476a-8c5b-85774b8be5bc-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.595947 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff686c4d-3ada-476a-8c5b-85774b8be5bc-kube-api-access-8n6h5" (OuterVolumeSpecName: "kube-api-access-8n6h5") pod "ff686c4d-3ada-476a-8c5b-85774b8be5bc" (UID: "ff686c4d-3ada-476a-8c5b-85774b8be5bc"). InnerVolumeSpecName "kube-api-access-8n6h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.679145 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff686c4d-3ada-476a-8c5b-85774b8be5bc" (UID: "ff686c4d-3ada-476a-8c5b-85774b8be5bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.686368 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-config-data" (OuterVolumeSpecName: "config-data") pod "ff686c4d-3ada-476a-8c5b-85774b8be5bc" (UID: "ff686c4d-3ada-476a-8c5b-85774b8be5bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.691881 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.691908 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n6h5\" (UniqueName: \"kubernetes.io/projected/ff686c4d-3ada-476a-8c5b-85774b8be5bc-kube-api-access-8n6h5\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.691920 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff686c4d-3ada-476a-8c5b-85774b8be5bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.826742 4679 generic.go:334] "Generic (PLEG): container finished" podID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerID="0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966" exitCode=0 Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.826801 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.827041 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff686c4d-3ada-476a-8c5b-85774b8be5bc","Type":"ContainerDied","Data":"0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966"} Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.827153 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff686c4d-3ada-476a-8c5b-85774b8be5bc","Type":"ContainerDied","Data":"a52241d780bab9e7fafee846e71810db3ec89f849bb03184e21028c7043e8dc3"} Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.827182 4679 scope.go:117] "RemoveContainer" containerID="0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.829222 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerStarted","Data":"ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af"} Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.859122 4679 scope.go:117] "RemoveContainer" containerID="5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.869410 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.886936 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.897803 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:29 crc kubenswrapper[4679]: E1202 10:41:29.898129 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-api" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.898146 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-api" Dec 02 10:41:29 crc kubenswrapper[4679]: E1202 10:41:29.898183 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-log" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.898190 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-log" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.898374 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-log" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.898404 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" containerName="nova-api-api" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.900847 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.908476 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.908818 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.911244 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.926499 4679 scope.go:117] "RemoveContainer" containerID="0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966" Dec 02 10:41:29 crc kubenswrapper[4679]: E1202 10:41:29.927962 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966\": container with ID starting with 0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966 not found: ID does not exist" containerID="0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.927995 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966"} err="failed to get container status \"0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966\": rpc error: code = NotFound desc = could not find container \"0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966\": container with ID starting with 0605a5afe0a3c7ac384aa3cd2789bce850bfb1bef4c15dc5a7757b2d63e5f966 not found: ID does not exist" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.928021 4679 scope.go:117] "RemoveContainer" containerID="5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e" Dec 02 10:41:29 crc kubenswrapper[4679]: E1202 10:41:29.928467 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e\": container with ID starting with 5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e not found: ID does not exist" containerID="5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.928508 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e"} err="failed to get container status \"5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e\": rpc error: code = NotFound desc = could not find container \"5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e\": container with ID starting with 5454614444b5a3446e129dba0e5801b3d1d8577c413cdb04c3f2bbc7cbc7147e not found: ID does not exist" Dec 02 10:41:29 crc kubenswrapper[4679]: I1202 10:41:29.956415 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.102532 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.102677 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-config-data\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.102780 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5359e3ba-1275-4858-b0d0-8e49e96ab30d-logs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.102832 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndwpm\" (UniqueName: \"kubernetes.io/projected/5359e3ba-1275-4858-b0d0-8e49e96ab30d-kube-api-access-ndwpm\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.102940 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-public-tls-certs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.102969 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.205165 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-config-data\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.205209 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5359e3ba-1275-4858-b0d0-8e49e96ab30d-logs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.205233 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndwpm\" (UniqueName: \"kubernetes.io/projected/5359e3ba-1275-4858-b0d0-8e49e96ab30d-kube-api-access-ndwpm\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.205289 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-public-tls-certs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.205391 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.205454 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.206755 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5359e3ba-1275-4858-b0d0-8e49e96ab30d-logs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.210124 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.210148 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-public-tls-certs\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.213795 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-config-data\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.216293 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.225145 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndwpm\" (UniqueName: \"kubernetes.io/projected/5359e3ba-1275-4858-b0d0-8e49e96ab30d-kube-api-access-ndwpm\") pod \"nova-api-0\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.232319 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.708612 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.839423 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5359e3ba-1275-4858-b0d0-8e49e96ab30d","Type":"ContainerStarted","Data":"ed55287a0fc98581e2574563382e59378769113d1be7244324012f7aef79ddce"} Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.846743 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerStarted","Data":"e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829"} Dec 02 10:41:30 crc kubenswrapper[4679]: I1202 10:41:30.919474 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff686c4d-3ada-476a-8c5b-85774b8be5bc" path="/var/lib/kubelet/pods/ff686c4d-3ada-476a-8c5b-85774b8be5bc/volumes" Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.379682 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.401552 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.860117 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5359e3ba-1275-4858-b0d0-8e49e96ab30d","Type":"ContainerStarted","Data":"d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094"} Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.860438 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5359e3ba-1275-4858-b0d0-8e49e96ab30d","Type":"ContainerStarted","Data":"587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc"} Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.868379 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerStarted","Data":"50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a"} Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.884394 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.88437221 podStartE2EDuration="2.88437221s" podCreationTimestamp="2025-12-02 10:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:31.882042483 +0000 UTC m=+1345.212181353" watchObservedRunningTime="2025-12-02 10:41:31.88437221 +0000 UTC m=+1345.214511070" Dec 02 10:41:31 crc kubenswrapper[4679]: I1202 10:41:31.903403 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.060849 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gg8gl"] Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.062096 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.064654 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.065123 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.074240 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gg8gl"] Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.240203 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-scripts\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.240272 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fq4l\" (UniqueName: \"kubernetes.io/projected/d4804674-9946-4546-aca6-f12ca930cbf8-kube-api-access-5fq4l\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.240632 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-config-data\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.240712 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.342831 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-config-data\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.342889 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.342919 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-scripts\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.342944 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fq4l\" (UniqueName: \"kubernetes.io/projected/d4804674-9946-4546-aca6-f12ca930cbf8-kube-api-access-5fq4l\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.357542 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.357818 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-scripts\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.364597 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-config-data\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.374524 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fq4l\" (UniqueName: \"kubernetes.io/projected/d4804674-9946-4546-aca6-f12ca930cbf8-kube-api-access-5fq4l\") pod \"nova-cell1-cell-mapping-gg8gl\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.414239 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.878520 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerStarted","Data":"5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63"} Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.878799 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="proxy-httpd" containerID="cri-o://5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63" gracePeriod=30 Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.878845 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="sg-core" containerID="cri-o://50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a" gracePeriod=30 Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.878796 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-central-agent" containerID="cri-o://ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af" gracePeriod=30 Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.878874 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-notification-agent" containerID="cri-o://e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829" gracePeriod=30 Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.878967 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.916643 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.095117061 podStartE2EDuration="6.916629104s" podCreationTimestamp="2025-12-02 10:41:26 +0000 UTC" firstStartedPulling="2025-12-02 10:41:27.59414239 +0000 UTC m=+1340.924281250" lastFinishedPulling="2025-12-02 10:41:32.415654433 +0000 UTC m=+1345.745793293" observedRunningTime="2025-12-02 10:41:32.913101573 +0000 UTC m=+1346.243240433" watchObservedRunningTime="2025-12-02 10:41:32.916629104 +0000 UTC m=+1346.246767964" Dec 02 10:41:32 crc kubenswrapper[4679]: I1202 10:41:32.921368 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gg8gl"] Dec 02 10:41:32 crc kubenswrapper[4679]: W1202 10:41:32.926899 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4804674_9946_4546_aca6_f12ca930cbf8.slice/crio-a9c786cc72ec329bccd69c75b3a7b828224818b85308cd5eadd04395677293ba WatchSource:0}: Error finding container a9c786cc72ec329bccd69c75b3a7b828224818b85308cd5eadd04395677293ba: Status 404 returned error can't find the container with id a9c786cc72ec329bccd69c75b3a7b828224818b85308cd5eadd04395677293ba Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.291471 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.387252 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-drnpf"] Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.387484 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerName="dnsmasq-dns" containerID="cri-o://adc0b6d07a9e004ccd5957f680457754553a5174e4b24399f291e77d4c49a63b" gracePeriod=10 Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.889071 4679 generic.go:334] "Generic (PLEG): container finished" podID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerID="adc0b6d07a9e004ccd5957f680457754553a5174e4b24399f291e77d4c49a63b" exitCode=0 Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.889152 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" event={"ID":"5d16bd4a-4e68-43d9-8b62-98825176cd75","Type":"ContainerDied","Data":"adc0b6d07a9e004ccd5957f680457754553a5174e4b24399f291e77d4c49a63b"} Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.890703 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gg8gl" event={"ID":"d4804674-9946-4546-aca6-f12ca930cbf8","Type":"ContainerStarted","Data":"375d036778fc2db2dc4e0c6120ae42fe941461104c98914afa3117759d602ce8"} Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.890758 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gg8gl" event={"ID":"d4804674-9946-4546-aca6-f12ca930cbf8","Type":"ContainerStarted","Data":"a9c786cc72ec329bccd69c75b3a7b828224818b85308cd5eadd04395677293ba"} Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.894029 4679 generic.go:334] "Generic (PLEG): container finished" podID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerID="5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63" exitCode=0 Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.894057 4679 generic.go:334] "Generic (PLEG): container finished" podID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerID="50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a" exitCode=2 Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.894065 4679 generic.go:334] "Generic (PLEG): container finished" podID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerID="e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829" exitCode=0 Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.894092 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerDied","Data":"5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63"} Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.894114 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerDied","Data":"50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a"} Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.894124 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerDied","Data":"e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829"} Dec 02 10:41:33 crc kubenswrapper[4679]: I1202 10:41:33.912143 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gg8gl" podStartSLOduration=1.9121268809999998 podStartE2EDuration="1.912126881s" podCreationTimestamp="2025-12-02 10:41:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:33.903803051 +0000 UTC m=+1347.233941901" watchObservedRunningTime="2025-12-02 10:41:33.912126881 +0000 UTC m=+1347.242265741" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.363282 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.480577 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-svc\") pod \"5d16bd4a-4e68-43d9-8b62-98825176cd75\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.480692 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5bx4\" (UniqueName: \"kubernetes.io/projected/5d16bd4a-4e68-43d9-8b62-98825176cd75-kube-api-access-n5bx4\") pod \"5d16bd4a-4e68-43d9-8b62-98825176cd75\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.480722 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-nb\") pod \"5d16bd4a-4e68-43d9-8b62-98825176cd75\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.480776 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-sb\") pod \"5d16bd4a-4e68-43d9-8b62-98825176cd75\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.480861 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-config\") pod \"5d16bd4a-4e68-43d9-8b62-98825176cd75\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.480926 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-swift-storage-0\") pod \"5d16bd4a-4e68-43d9-8b62-98825176cd75\" (UID: \"5d16bd4a-4e68-43d9-8b62-98825176cd75\") " Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.502545 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d16bd4a-4e68-43d9-8b62-98825176cd75-kube-api-access-n5bx4" (OuterVolumeSpecName: "kube-api-access-n5bx4") pod "5d16bd4a-4e68-43d9-8b62-98825176cd75" (UID: "5d16bd4a-4e68-43d9-8b62-98825176cd75"). InnerVolumeSpecName "kube-api-access-n5bx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.536596 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5d16bd4a-4e68-43d9-8b62-98825176cd75" (UID: "5d16bd4a-4e68-43d9-8b62-98825176cd75"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.540685 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-config" (OuterVolumeSpecName: "config") pod "5d16bd4a-4e68-43d9-8b62-98825176cd75" (UID: "5d16bd4a-4e68-43d9-8b62-98825176cd75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.542042 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5d16bd4a-4e68-43d9-8b62-98825176cd75" (UID: "5d16bd4a-4e68-43d9-8b62-98825176cd75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.544382 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5d16bd4a-4e68-43d9-8b62-98825176cd75" (UID: "5d16bd4a-4e68-43d9-8b62-98825176cd75"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.561317 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d16bd4a-4e68-43d9-8b62-98825176cd75" (UID: "5d16bd4a-4e68-43d9-8b62-98825176cd75"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.582762 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.582800 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.582812 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.582821 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5bx4\" (UniqueName: \"kubernetes.io/projected/5d16bd4a-4e68-43d9-8b62-98825176cd75-kube-api-access-n5bx4\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.582829 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.582837 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d16bd4a-4e68-43d9-8b62-98825176cd75-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.920112 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.934799 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-drnpf" event={"ID":"5d16bd4a-4e68-43d9-8b62-98825176cd75","Type":"ContainerDied","Data":"ef13fcb044157e7d8ec6777ca2bd382261bfc6365949e7d3bebc67c5e3c719fe"} Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.934855 4679 scope.go:117] "RemoveContainer" containerID="adc0b6d07a9e004ccd5957f680457754553a5174e4b24399f291e77d4c49a63b" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.967085 4679 scope.go:117] "RemoveContainer" containerID="39f33088a3a25c2c1cd3e8c91ee418ea04d4016736a7296fe08ccc7b241ed237" Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.977630 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-drnpf"] Dec 02 10:41:34 crc kubenswrapper[4679]: I1202 10:41:34.986067 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-drnpf"] Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.183424 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.297946 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-scripts\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298034 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-log-httpd\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298088 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-combined-ca-bundle\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298108 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d284z\" (UniqueName: \"kubernetes.io/projected/24c4ed59-db75-4d18-9889-bfea9ca275b7-kube-api-access-d284z\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298137 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-ceilometer-tls-certs\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298166 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-config-data\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298203 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-run-httpd\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.298351 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-sg-core-conf-yaml\") pod \"24c4ed59-db75-4d18-9889-bfea9ca275b7\" (UID: \"24c4ed59-db75-4d18-9889-bfea9ca275b7\") " Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.299298 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.299559 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.303563 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24c4ed59-db75-4d18-9889-bfea9ca275b7-kube-api-access-d284z" (OuterVolumeSpecName: "kube-api-access-d284z") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "kube-api-access-d284z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.303627 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-scripts" (OuterVolumeSpecName: "scripts") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.323713 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.346809 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.365544 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.386826 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-config-data" (OuterVolumeSpecName: "config-data") pod "24c4ed59-db75-4d18-9889-bfea9ca275b7" (UID: "24c4ed59-db75-4d18-9889-bfea9ca275b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401465 4679 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401515 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401531 4679 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401548 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401569 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d284z\" (UniqueName: \"kubernetes.io/projected/24c4ed59-db75-4d18-9889-bfea9ca275b7-kube-api-access-d284z\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401590 4679 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401606 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c4ed59-db75-4d18-9889-bfea9ca275b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.401625 4679 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24c4ed59-db75-4d18-9889-bfea9ca275b7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.927541 4679 generic.go:334] "Generic (PLEG): container finished" podID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerID="ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af" exitCode=0 Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.927610 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.927640 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerDied","Data":"ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af"} Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.928013 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24c4ed59-db75-4d18-9889-bfea9ca275b7","Type":"ContainerDied","Data":"1d126fd2fef35eb431c032e6e1982dd6c58ddedbdad79efcb39adbb174750927"} Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.928055 4679 scope.go:117] "RemoveContainer" containerID="5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.952165 4679 scope.go:117] "RemoveContainer" containerID="50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.966812 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.979442 4679 scope.go:117] "RemoveContainer" containerID="e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829" Dec 02 10:41:35 crc kubenswrapper[4679]: I1202 10:41:35.986205 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002381 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.002791 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-central-agent" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002807 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-central-agent" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.002822 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerName="init" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002829 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerName="init" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.002840 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-notification-agent" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002848 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-notification-agent" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.002859 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerName="dnsmasq-dns" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002866 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerName="dnsmasq-dns" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.002875 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="proxy-httpd" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002880 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="proxy-httpd" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.002907 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="sg-core" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.002913 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="sg-core" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.003082 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="sg-core" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.003093 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-notification-agent" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.003107 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" containerName="dnsmasq-dns" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.003120 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="ceilometer-central-agent" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.003139 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" containerName="proxy-httpd" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.004799 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.006599 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.006820 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.007105 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.010162 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.022714 4679 scope.go:117] "RemoveContainer" containerID="ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.079515 4679 scope.go:117] "RemoveContainer" containerID="5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.080439 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63\": container with ID starting with 5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63 not found: ID does not exist" containerID="5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.080645 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63"} err="failed to get container status \"5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63\": rpc error: code = NotFound desc = could not find container \"5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63\": container with ID starting with 5bd98cae3a50efafe29a31033e9f28d34e934df9832fc9b17b2831f27a002d63 not found: ID does not exist" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.080819 4679 scope.go:117] "RemoveContainer" containerID="50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.081329 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a\": container with ID starting with 50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a not found: ID does not exist" containerID="50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.081433 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a"} err="failed to get container status \"50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a\": rpc error: code = NotFound desc = could not find container \"50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a\": container with ID starting with 50d29240f4293c559e861a16474c475e486fd48b03a78b9617a4419f8165596a not found: ID does not exist" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.081527 4679 scope.go:117] "RemoveContainer" containerID="e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.081805 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829\": container with ID starting with e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829 not found: ID does not exist" containerID="e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.081885 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829"} err="failed to get container status \"e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829\": rpc error: code = NotFound desc = could not find container \"e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829\": container with ID starting with e1fce8442b4929c10e8bb73a9abd3ab1ee95ac98697767c1e8bb026e92dfa829 not found: ID does not exist" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.081949 4679 scope.go:117] "RemoveContainer" containerID="ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af" Dec 02 10:41:36 crc kubenswrapper[4679]: E1202 10:41:36.082197 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af\": container with ID starting with ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af not found: ID does not exist" containerID="ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.082278 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af"} err="failed to get container status \"ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af\": rpc error: code = NotFound desc = could not find container \"ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af\": container with ID starting with ce7c2e411c0775c597d55a488b10c102789adaa95129db56f6f7af804968f7af not found: ID does not exist" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114170 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-scripts\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114237 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114361 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17b54231-dc1b-4ccd-8679-b9a697615dbb-log-httpd\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114384 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbxvh\" (UniqueName: \"kubernetes.io/projected/17b54231-dc1b-4ccd-8679-b9a697615dbb-kube-api-access-zbxvh\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114399 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114575 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-config-data\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114676 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.114759 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17b54231-dc1b-4ccd-8679-b9a697615dbb-run-httpd\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217434 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17b54231-dc1b-4ccd-8679-b9a697615dbb-log-httpd\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217497 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbxvh\" (UniqueName: \"kubernetes.io/projected/17b54231-dc1b-4ccd-8679-b9a697615dbb-kube-api-access-zbxvh\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217514 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217574 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-config-data\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217594 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217636 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17b54231-dc1b-4ccd-8679-b9a697615dbb-run-httpd\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217699 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-scripts\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.217726 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.218403 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17b54231-dc1b-4ccd-8679-b9a697615dbb-run-httpd\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.221476 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17b54231-dc1b-4ccd-8679-b9a697615dbb-log-httpd\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.222067 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.222184 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-scripts\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.223022 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.225049 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-config-data\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.227899 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b54231-dc1b-4ccd-8679-b9a697615dbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.234925 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbxvh\" (UniqueName: \"kubernetes.io/projected/17b54231-dc1b-4ccd-8679-b9a697615dbb-kube-api-access-zbxvh\") pod \"ceilometer-0\" (UID: \"17b54231-dc1b-4ccd-8679-b9a697615dbb\") " pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.379546 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.843057 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.927559 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24c4ed59-db75-4d18-9889-bfea9ca275b7" path="/var/lib/kubelet/pods/24c4ed59-db75-4d18-9889-bfea9ca275b7/volumes" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.928550 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d16bd4a-4e68-43d9-8b62-98825176cd75" path="/var/lib/kubelet/pods/5d16bd4a-4e68-43d9-8b62-98825176cd75/volumes" Dec 02 10:41:36 crc kubenswrapper[4679]: I1202 10:41:36.942723 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17b54231-dc1b-4ccd-8679-b9a697615dbb","Type":"ContainerStarted","Data":"0ed0de836843df96826059dccb2d6403e3a0f15fd1bcf8f912cef3a2a96287e5"} Dec 02 10:41:38 crc kubenswrapper[4679]: I1202 10:41:38.960224 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17b54231-dc1b-4ccd-8679-b9a697615dbb","Type":"ContainerStarted","Data":"4e0f4bf7bbc087274e4b56764e97f90fa1c75fd144205680b156949b54eb08f5"} Dec 02 10:41:38 crc kubenswrapper[4679]: I1202 10:41:38.960797 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17b54231-dc1b-4ccd-8679-b9a697615dbb","Type":"ContainerStarted","Data":"d9f677032cec405121ffe611b81e5a58a59a1fdf2f7e441604b6052094825b94"} Dec 02 10:41:38 crc kubenswrapper[4679]: I1202 10:41:38.962472 4679 generic.go:334] "Generic (PLEG): container finished" podID="d4804674-9946-4546-aca6-f12ca930cbf8" containerID="375d036778fc2db2dc4e0c6120ae42fe941461104c98914afa3117759d602ce8" exitCode=0 Dec 02 10:41:38 crc kubenswrapper[4679]: I1202 10:41:38.962535 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gg8gl" event={"ID":"d4804674-9946-4546-aca6-f12ca930cbf8","Type":"ContainerDied","Data":"375d036778fc2db2dc4e0c6120ae42fe941461104c98914afa3117759d602ce8"} Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.232747 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.233213 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.399192 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.531658 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fq4l\" (UniqueName: \"kubernetes.io/projected/d4804674-9946-4546-aca6-f12ca930cbf8-kube-api-access-5fq4l\") pod \"d4804674-9946-4546-aca6-f12ca930cbf8\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.531754 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-combined-ca-bundle\") pod \"d4804674-9946-4546-aca6-f12ca930cbf8\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.531809 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-config-data\") pod \"d4804674-9946-4546-aca6-f12ca930cbf8\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.531834 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-scripts\") pod \"d4804674-9946-4546-aca6-f12ca930cbf8\" (UID: \"d4804674-9946-4546-aca6-f12ca930cbf8\") " Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.537639 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-scripts" (OuterVolumeSpecName: "scripts") pod "d4804674-9946-4546-aca6-f12ca930cbf8" (UID: "d4804674-9946-4546-aca6-f12ca930cbf8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.537816 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4804674-9946-4546-aca6-f12ca930cbf8-kube-api-access-5fq4l" (OuterVolumeSpecName: "kube-api-access-5fq4l") pod "d4804674-9946-4546-aca6-f12ca930cbf8" (UID: "d4804674-9946-4546-aca6-f12ca930cbf8"). InnerVolumeSpecName "kube-api-access-5fq4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.559559 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-config-data" (OuterVolumeSpecName: "config-data") pod "d4804674-9946-4546-aca6-f12ca930cbf8" (UID: "d4804674-9946-4546-aca6-f12ca930cbf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.563672 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4804674-9946-4546-aca6-f12ca930cbf8" (UID: "d4804674-9946-4546-aca6-f12ca930cbf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.634152 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fq4l\" (UniqueName: \"kubernetes.io/projected/d4804674-9946-4546-aca6-f12ca930cbf8-kube-api-access-5fq4l\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.634197 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.634208 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.634216 4679 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4804674-9946-4546-aca6-f12ca930cbf8-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.983493 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17b54231-dc1b-4ccd-8679-b9a697615dbb","Type":"ContainerStarted","Data":"282d89ee14be405b73499a469ecf102d869b578af670a7f750ad1798489fbf06"} Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.985276 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gg8gl" event={"ID":"d4804674-9946-4546-aca6-f12ca930cbf8","Type":"ContainerDied","Data":"a9c786cc72ec329bccd69c75b3a7b828224818b85308cd5eadd04395677293ba"} Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.985314 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c786cc72ec329bccd69c75b3a7b828224818b85308cd5eadd04395677293ba" Dec 02 10:41:40 crc kubenswrapper[4679]: I1202 10:41:40.985362 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gg8gl" Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.132018 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.132272 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-log" containerID="cri-o://587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc" gracePeriod=30 Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.132340 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-api" containerID="cri-o://d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094" gracePeriod=30 Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.146573 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.146815 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4c7bbbe3-232a-41b2-9378-848755f2d1e6" containerName="nova-scheduler-scheduler" containerID="cri-o://5e9e6d52b8d0c88e455191d96c76b4befb97c7f8b7ffb0ffee8623f646d8a7d5" gracePeriod=30 Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.147297 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": EOF" Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.147326 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": EOF" Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.164748 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.165013 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-log" containerID="cri-o://ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf" gracePeriod=30 Dec 02 10:41:41 crc kubenswrapper[4679]: I1202 10:41:41.165532 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-metadata" containerID="cri-o://bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf" gracePeriod=30 Dec 02 10:41:42 crc kubenswrapper[4679]: I1202 10:41:42.002458 4679 generic.go:334] "Generic (PLEG): container finished" podID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerID="587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc" exitCode=143 Dec 02 10:41:42 crc kubenswrapper[4679]: I1202 10:41:42.002552 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5359e3ba-1275-4858-b0d0-8e49e96ab30d","Type":"ContainerDied","Data":"587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc"} Dec 02 10:41:42 crc kubenswrapper[4679]: I1202 10:41:42.006150 4679 generic.go:334] "Generic (PLEG): container finished" podID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerID="ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf" exitCode=143 Dec 02 10:41:42 crc kubenswrapper[4679]: I1202 10:41:42.006205 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4bc7924b-f6dd-4fb9-bf1c-64af7274a161","Type":"ContainerDied","Data":"ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf"} Dec 02 10:41:43 crc kubenswrapper[4679]: I1202 10:41:43.016812 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17b54231-dc1b-4ccd-8679-b9a697615dbb","Type":"ContainerStarted","Data":"7ba7fe703dd71ca7855d9d74da59538a49aca23c76acc6549125e26db161df29"} Dec 02 10:41:43 crc kubenswrapper[4679]: I1202 10:41:43.017403 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 10:41:43 crc kubenswrapper[4679]: I1202 10:41:43.047752 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.903969564 podStartE2EDuration="8.04773672s" podCreationTimestamp="2025-12-02 10:41:35 +0000 UTC" firstStartedPulling="2025-12-02 10:41:36.846402712 +0000 UTC m=+1350.176541572" lastFinishedPulling="2025-12-02 10:41:41.990169868 +0000 UTC m=+1355.320308728" observedRunningTime="2025-12-02 10:41:43.036442405 +0000 UTC m=+1356.366581295" watchObservedRunningTime="2025-12-02 10:41:43.04773672 +0000 UTC m=+1356.377875580" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.028903 4679 generic.go:334] "Generic (PLEG): container finished" podID="4c7bbbe3-232a-41b2-9378-848755f2d1e6" containerID="5e9e6d52b8d0c88e455191d96c76b4befb97c7f8b7ffb0ffee8623f646d8a7d5" exitCode=0 Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.029009 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4c7bbbe3-232a-41b2-9378-848755f2d1e6","Type":"ContainerDied","Data":"5e9e6d52b8d0c88e455191d96c76b4befb97c7f8b7ffb0ffee8623f646d8a7d5"} Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.322206 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.414097 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f8cg\" (UniqueName: \"kubernetes.io/projected/4c7bbbe3-232a-41b2-9378-848755f2d1e6-kube-api-access-4f8cg\") pod \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.414239 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-combined-ca-bundle\") pod \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.414267 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-config-data\") pod \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\" (UID: \"4c7bbbe3-232a-41b2-9378-848755f2d1e6\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.423629 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7bbbe3-232a-41b2-9378-848755f2d1e6-kube-api-access-4f8cg" (OuterVolumeSpecName: "kube-api-access-4f8cg") pod "4c7bbbe3-232a-41b2-9378-848755f2d1e6" (UID: "4c7bbbe3-232a-41b2-9378-848755f2d1e6"). InnerVolumeSpecName "kube-api-access-4f8cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.428450 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:40612->10.217.0.191:8775: read: connection reset by peer" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.428822 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:40608->10.217.0.191:8775: read: connection reset by peer" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.453752 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c7bbbe3-232a-41b2-9378-848755f2d1e6" (UID: "4c7bbbe3-232a-41b2-9378-848755f2d1e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.455919 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-config-data" (OuterVolumeSpecName: "config-data") pod "4c7bbbe3-232a-41b2-9378-848755f2d1e6" (UID: "4c7bbbe3-232a-41b2-9378-848755f2d1e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.525458 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f8cg\" (UniqueName: \"kubernetes.io/projected/4c7bbbe3-232a-41b2-9378-848755f2d1e6-kube-api-access-4f8cg\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.525508 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.525521 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c7bbbe3-232a-41b2-9378-848755f2d1e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.834699 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.931928 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-nova-metadata-tls-certs\") pod \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.932000 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-combined-ca-bundle\") pod \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.932121 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-config-data\") pod \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.932163 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd7q2\" (UniqueName: \"kubernetes.io/projected/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-kube-api-access-gd7q2\") pod \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.932380 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-logs\") pod \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\" (UID: \"4bc7924b-f6dd-4fb9-bf1c-64af7274a161\") " Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.933386 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-logs" (OuterVolumeSpecName: "logs") pod "4bc7924b-f6dd-4fb9-bf1c-64af7274a161" (UID: "4bc7924b-f6dd-4fb9-bf1c-64af7274a161"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.938519 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-kube-api-access-gd7q2" (OuterVolumeSpecName: "kube-api-access-gd7q2") pod "4bc7924b-f6dd-4fb9-bf1c-64af7274a161" (UID: "4bc7924b-f6dd-4fb9-bf1c-64af7274a161"). InnerVolumeSpecName "kube-api-access-gd7q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.969787 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-config-data" (OuterVolumeSpecName: "config-data") pod "4bc7924b-f6dd-4fb9-bf1c-64af7274a161" (UID: "4bc7924b-f6dd-4fb9-bf1c-64af7274a161"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:44 crc kubenswrapper[4679]: I1202 10:41:44.970248 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bc7924b-f6dd-4fb9-bf1c-64af7274a161" (UID: "4bc7924b-f6dd-4fb9-bf1c-64af7274a161"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.013331 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4bc7924b-f6dd-4fb9-bf1c-64af7274a161" (UID: "4bc7924b-f6dd-4fb9-bf1c-64af7274a161"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.035819 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.035863 4679 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.035877 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.035890 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.035903 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd7q2\" (UniqueName: \"kubernetes.io/projected/4bc7924b-f6dd-4fb9-bf1c-64af7274a161-kube-api-access-gd7q2\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.051816 4679 generic.go:334] "Generic (PLEG): container finished" podID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerID="bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf" exitCode=0 Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.051889 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4bc7924b-f6dd-4fb9-bf1c-64af7274a161","Type":"ContainerDied","Data":"bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf"} Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.051922 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4bc7924b-f6dd-4fb9-bf1c-64af7274a161","Type":"ContainerDied","Data":"44018de466426ced3419c9378ebb562670aa7438620f753bc45ac75c15e5b82a"} Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.051942 4679 scope.go:117] "RemoveContainer" containerID="bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.052092 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.066353 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4c7bbbe3-232a-41b2-9378-848755f2d1e6","Type":"ContainerDied","Data":"a8674757fc33c9eb00f02c99c4539b155d88db691351c9636a69fb4bac28fc35"} Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.066431 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.089887 4679 scope.go:117] "RemoveContainer" containerID="ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.099884 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.124459 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.125077 4679 scope.go:117] "RemoveContainer" containerID="bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf" Dec 02 10:41:45 crc kubenswrapper[4679]: E1202 10:41:45.129183 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf\": container with ID starting with bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf not found: ID does not exist" containerID="bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.129240 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf"} err="failed to get container status \"bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf\": rpc error: code = NotFound desc = could not find container \"bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf\": container with ID starting with bae0d068e8003ccbb6418e961bd8604b5c0a4af80f3bd80ac3e96c52861d34cf not found: ID does not exist" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.129272 4679 scope.go:117] "RemoveContainer" containerID="ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf" Dec 02 10:41:45 crc kubenswrapper[4679]: E1202 10:41:45.131900 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf\": container with ID starting with ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf not found: ID does not exist" containerID="ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.131957 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf"} err="failed to get container status \"ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf\": rpc error: code = NotFound desc = could not find container \"ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf\": container with ID starting with ef0cb8146de2824b8694e261235361efda031ba76068d1137a0cbafdd660c5cf not found: ID does not exist" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.131990 4679 scope.go:117] "RemoveContainer" containerID="5e9e6d52b8d0c88e455191d96c76b4befb97c7f8b7ffb0ffee8623f646d8a7d5" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.147463 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.166377 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.182996 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: E1202 10:41:45.183580 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-log" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183596 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-log" Dec 02 10:41:45 crc kubenswrapper[4679]: E1202 10:41:45.183623 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-metadata" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183631 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-metadata" Dec 02 10:41:45 crc kubenswrapper[4679]: E1202 10:41:45.183661 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7bbbe3-232a-41b2-9378-848755f2d1e6" containerName="nova-scheduler-scheduler" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183669 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7bbbe3-232a-41b2-9378-848755f2d1e6" containerName="nova-scheduler-scheduler" Dec 02 10:41:45 crc kubenswrapper[4679]: E1202 10:41:45.183678 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4804674-9946-4546-aca6-f12ca930cbf8" containerName="nova-manage" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183685 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4804674-9946-4546-aca6-f12ca930cbf8" containerName="nova-manage" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183906 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4804674-9946-4546-aca6-f12ca930cbf8" containerName="nova-manage" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183938 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7bbbe3-232a-41b2-9378-848755f2d1e6" containerName="nova-scheduler-scheduler" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183952 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-metadata" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.183970 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" containerName="nova-metadata-log" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.185257 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.187116 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.187464 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.193983 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.196818 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.199850 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.207379 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.217282 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341144 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-config-data\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341238 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pngf\" (UniqueName: \"kubernetes.io/projected/d731d782-b47c-4e03-8191-47ae5d52a555-kube-api-access-7pngf\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341430 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341534 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341592 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d731d782-b47c-4e03-8191-47ae5d52a555-config-data\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341692 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d049f08-8029-4ef7-9383-34925ba3d48b-logs\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341747 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d731d782-b47c-4e03-8191-47ae5d52a555-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.341858 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6hjk\" (UniqueName: \"kubernetes.io/projected/8d049f08-8029-4ef7-9383-34925ba3d48b-kube-api-access-j6hjk\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444205 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6hjk\" (UniqueName: \"kubernetes.io/projected/8d049f08-8029-4ef7-9383-34925ba3d48b-kube-api-access-j6hjk\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444360 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-config-data\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444478 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pngf\" (UniqueName: \"kubernetes.io/projected/d731d782-b47c-4e03-8191-47ae5d52a555-kube-api-access-7pngf\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444531 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444576 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444623 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d731d782-b47c-4e03-8191-47ae5d52a555-config-data\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444682 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d049f08-8029-4ef7-9383-34925ba3d48b-logs\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.444720 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d731d782-b47c-4e03-8191-47ae5d52a555-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.445479 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d049f08-8029-4ef7-9383-34925ba3d48b-logs\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.451004 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.451298 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d731d782-b47c-4e03-8191-47ae5d52a555-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.452636 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d731d782-b47c-4e03-8191-47ae5d52a555-config-data\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.457560 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.458011 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d049f08-8029-4ef7-9383-34925ba3d48b-config-data\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.470999 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pngf\" (UniqueName: \"kubernetes.io/projected/d731d782-b47c-4e03-8191-47ae5d52a555-kube-api-access-7pngf\") pod \"nova-scheduler-0\" (UID: \"d731d782-b47c-4e03-8191-47ae5d52a555\") " pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.480992 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6hjk\" (UniqueName: \"kubernetes.io/projected/8d049f08-8029-4ef7-9383-34925ba3d48b-kube-api-access-j6hjk\") pod \"nova-metadata-0\" (UID: \"8d049f08-8029-4ef7-9383-34925ba3d48b\") " pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.501881 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.520539 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 10:41:45 crc kubenswrapper[4679]: I1202 10:41:45.913402 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.079138 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d731d782-b47c-4e03-8191-47ae5d52a555","Type":"ContainerStarted","Data":"e5b726dd0a28829959b70a24d2736a140ac98936f1338f43f83006b9da086d20"} Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.194925 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.914733 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.918557 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc7924b-f6dd-4fb9-bf1c-64af7274a161" path="/var/lib/kubelet/pods/4bc7924b-f6dd-4fb9-bf1c-64af7274a161/volumes" Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.919123 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7bbbe3-232a-41b2-9378-848755f2d1e6" path="/var/lib/kubelet/pods/4c7bbbe3-232a-41b2-9378-848755f2d1e6/volumes" Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.980092 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-config-data\") pod \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.980532 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5359e3ba-1275-4858-b0d0-8e49e96ab30d-logs\") pod \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.980615 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-public-tls-certs\") pod \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.980667 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-combined-ca-bundle\") pod \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.980720 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-internal-tls-certs\") pod \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.980764 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndwpm\" (UniqueName: \"kubernetes.io/projected/5359e3ba-1275-4858-b0d0-8e49e96ab30d-kube-api-access-ndwpm\") pod \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\" (UID: \"5359e3ba-1275-4858-b0d0-8e49e96ab30d\") " Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.982531 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5359e3ba-1275-4858-b0d0-8e49e96ab30d-logs" (OuterVolumeSpecName: "logs") pod "5359e3ba-1275-4858-b0d0-8e49e96ab30d" (UID: "5359e3ba-1275-4858-b0d0-8e49e96ab30d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:41:46 crc kubenswrapper[4679]: I1202 10:41:46.989023 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5359e3ba-1275-4858-b0d0-8e49e96ab30d-kube-api-access-ndwpm" (OuterVolumeSpecName: "kube-api-access-ndwpm") pod "5359e3ba-1275-4858-b0d0-8e49e96ab30d" (UID: "5359e3ba-1275-4858-b0d0-8e49e96ab30d"). InnerVolumeSpecName "kube-api-access-ndwpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.019000 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5359e3ba-1275-4858-b0d0-8e49e96ab30d" (UID: "5359e3ba-1275-4858-b0d0-8e49e96ab30d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.028142 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-config-data" (OuterVolumeSpecName: "config-data") pod "5359e3ba-1275-4858-b0d0-8e49e96ab30d" (UID: "5359e3ba-1275-4858-b0d0-8e49e96ab30d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.042280 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5359e3ba-1275-4858-b0d0-8e49e96ab30d" (UID: "5359e3ba-1275-4858-b0d0-8e49e96ab30d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.046725 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5359e3ba-1275-4858-b0d0-8e49e96ab30d" (UID: "5359e3ba-1275-4858-b0d0-8e49e96ab30d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.085517 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.085568 4679 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.085579 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndwpm\" (UniqueName: \"kubernetes.io/projected/5359e3ba-1275-4858-b0d0-8e49e96ab30d-kube-api-access-ndwpm\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.085592 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.085602 4679 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5359e3ba-1275-4858-b0d0-8e49e96ab30d-logs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.085611 4679 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5359e3ba-1275-4858-b0d0-8e49e96ab30d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.094074 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d049f08-8029-4ef7-9383-34925ba3d48b","Type":"ContainerStarted","Data":"049e3cfe2fb5cbaec4fcb8d0be34fa0e6aebe224fca85b4548517a5ee19919bf"} Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.094127 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d049f08-8029-4ef7-9383-34925ba3d48b","Type":"ContainerStarted","Data":"2e40baf93221c524933457f6f9d04b7b6fbd1d3465535e0ad7c3e2d452ebeab9"} Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.094137 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d049f08-8029-4ef7-9383-34925ba3d48b","Type":"ContainerStarted","Data":"a43b3f6818479b9e1cc1c8638f8ccb2c20347c98de565ce0ad1b3b3e01607ad6"} Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.098563 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d731d782-b47c-4e03-8191-47ae5d52a555","Type":"ContainerStarted","Data":"8de6a2432589555e5b912a5e37a5e234c2a21ef83a1f393a7a3537057fdc0364"} Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.101279 4679 generic.go:334] "Generic (PLEG): container finished" podID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerID="d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094" exitCode=0 Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.101336 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5359e3ba-1275-4858-b0d0-8e49e96ab30d","Type":"ContainerDied","Data":"d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094"} Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.101372 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5359e3ba-1275-4858-b0d0-8e49e96ab30d","Type":"ContainerDied","Data":"ed55287a0fc98581e2574563382e59378769113d1be7244324012f7aef79ddce"} Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.101389 4679 scope.go:117] "RemoveContainer" containerID="d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.101632 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.127370 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.127350266 podStartE2EDuration="2.127350266s" podCreationTimestamp="2025-12-02 10:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:47.111451649 +0000 UTC m=+1360.441590509" watchObservedRunningTime="2025-12-02 10:41:47.127350266 +0000 UTC m=+1360.457489126" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.139509 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.139493225 podStartE2EDuration="2.139493225s" podCreationTimestamp="2025-12-02 10:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:47.1351342 +0000 UTC m=+1360.465273060" watchObservedRunningTime="2025-12-02 10:41:47.139493225 +0000 UTC m=+1360.469632085" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.156063 4679 scope.go:117] "RemoveContainer" containerID="587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.162265 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.170929 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.194182 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:47 crc kubenswrapper[4679]: E1202 10:41:47.194655 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-api" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.194669 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-api" Dec 02 10:41:47 crc kubenswrapper[4679]: E1202 10:41:47.194688 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-log" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.194694 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-log" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.194868 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-log" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.194883 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" containerName="nova-api-api" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.195106 4679 scope.go:117] "RemoveContainer" containerID="d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094" Dec 02 10:41:47 crc kubenswrapper[4679]: E1202 10:41:47.195523 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094\": container with ID starting with d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094 not found: ID does not exist" containerID="d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.195560 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094"} err="failed to get container status \"d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094\": rpc error: code = NotFound desc = could not find container \"d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094\": container with ID starting with d2232bc079a2f44a430c9ccdc45e524e69bb76d31594c59421b2735e101a4094 not found: ID does not exist" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.195581 4679 scope.go:117] "RemoveContainer" containerID="587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc" Dec 02 10:41:47 crc kubenswrapper[4679]: E1202 10:41:47.195841 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc\": container with ID starting with 587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc not found: ID does not exist" containerID="587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.195869 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc"} err="failed to get container status \"587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc\": rpc error: code = NotFound desc = could not find container \"587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc\": container with ID starting with 587e2dda664889c469386cc8b7fa5678cf1e7654a61f5928535b63d83ddb9fdc not found: ID does not exist" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.195887 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.198253 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.198500 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.198604 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.220757 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.290210 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-internal-tls-certs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.290328 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85cv6\" (UniqueName: \"kubernetes.io/projected/81f22355-2ad7-4fd6-bc88-779cd704b038-kube-api-access-85cv6\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.290353 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f22355-2ad7-4fd6-bc88-779cd704b038-logs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.290383 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-public-tls-certs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.290608 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-config-data\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.290709 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.392397 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-config-data\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.392456 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.392493 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-internal-tls-certs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.392546 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85cv6\" (UniqueName: \"kubernetes.io/projected/81f22355-2ad7-4fd6-bc88-779cd704b038-kube-api-access-85cv6\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.392566 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f22355-2ad7-4fd6-bc88-779cd704b038-logs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.392592 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-public-tls-certs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.393289 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81f22355-2ad7-4fd6-bc88-779cd704b038-logs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.397432 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-config-data\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.397483 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-public-tls-certs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.397761 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-internal-tls-certs\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.406527 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81f22355-2ad7-4fd6-bc88-779cd704b038-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.410481 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85cv6\" (UniqueName: \"kubernetes.io/projected/81f22355-2ad7-4fd6-bc88-779cd704b038-kube-api-access-85cv6\") pod \"nova-api-0\" (UID: \"81f22355-2ad7-4fd6-bc88-779cd704b038\") " pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.521920 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 10:41:47 crc kubenswrapper[4679]: I1202 10:41:47.979495 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 10:41:47 crc kubenswrapper[4679]: W1202 10:41:47.980879 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81f22355_2ad7_4fd6_bc88_779cd704b038.slice/crio-4e85f5a0f5925b293a9acaebad65f05e1a97d6b51950f0709cbf4630ebc34b0d WatchSource:0}: Error finding container 4e85f5a0f5925b293a9acaebad65f05e1a97d6b51950f0709cbf4630ebc34b0d: Status 404 returned error can't find the container with id 4e85f5a0f5925b293a9acaebad65f05e1a97d6b51950f0709cbf4630ebc34b0d Dec 02 10:41:48 crc kubenswrapper[4679]: I1202 10:41:48.118240 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f22355-2ad7-4fd6-bc88-779cd704b038","Type":"ContainerStarted","Data":"4e85f5a0f5925b293a9acaebad65f05e1a97d6b51950f0709cbf4630ebc34b0d"} Dec 02 10:41:48 crc kubenswrapper[4679]: I1202 10:41:48.918483 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5359e3ba-1275-4858-b0d0-8e49e96ab30d" path="/var/lib/kubelet/pods/5359e3ba-1275-4858-b0d0-8e49e96ab30d/volumes" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.129620 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f22355-2ad7-4fd6-bc88-779cd704b038","Type":"ContainerStarted","Data":"9dad158d78b5d37ea2fe584ae9597b363539f8b42a8f29e8c70299cf8bfda098"} Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.130043 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81f22355-2ad7-4fd6-bc88-779cd704b038","Type":"ContainerStarted","Data":"2c52931caccc20132c748b7833d3b9463f43f2654427aad379b4816decd2d478"} Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.151764 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.151746 podStartE2EDuration="2.151746s" podCreationTimestamp="2025-12-02 10:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:41:49.147729175 +0000 UTC m=+1362.477868035" watchObservedRunningTime="2025-12-02 10:41:49.151746 +0000 UTC m=+1362.481884850" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.508238 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdmfv"] Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.510196 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.526734 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdmfv"] Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.635745 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-utilities\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.636113 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw6w8\" (UniqueName: \"kubernetes.io/projected/97c9b812-8677-4b11-96f5-b3537a7b9ab2-kube-api-access-lw6w8\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.636283 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-catalog-content\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.738101 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw6w8\" (UniqueName: \"kubernetes.io/projected/97c9b812-8677-4b11-96f5-b3537a7b9ab2-kube-api-access-lw6w8\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.738199 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-catalog-content\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.738288 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-utilities\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.738814 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-catalog-content\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.738846 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-utilities\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.756148 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw6w8\" (UniqueName: \"kubernetes.io/projected/97c9b812-8677-4b11-96f5-b3537a7b9ab2-kube-api-access-lw6w8\") pod \"redhat-operators-bdmfv\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:49 crc kubenswrapper[4679]: I1202 10:41:49.838083 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:50 crc kubenswrapper[4679]: I1202 10:41:50.322601 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdmfv"] Dec 02 10:41:50 crc kubenswrapper[4679]: W1202 10:41:50.328167 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97c9b812_8677_4b11_96f5_b3537a7b9ab2.slice/crio-bd536c448f687f0f54484473ed7a8f220c6610cb316b6c050a41bba568cbc5ac WatchSource:0}: Error finding container bd536c448f687f0f54484473ed7a8f220c6610cb316b6c050a41bba568cbc5ac: Status 404 returned error can't find the container with id bd536c448f687f0f54484473ed7a8f220c6610cb316b6c050a41bba568cbc5ac Dec 02 10:41:50 crc kubenswrapper[4679]: I1202 10:41:50.504019 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:41:50 crc kubenswrapper[4679]: I1202 10:41:50.504602 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 10:41:50 crc kubenswrapper[4679]: I1202 10:41:50.520703 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 10:41:51 crc kubenswrapper[4679]: I1202 10:41:51.148080 4679 generic.go:334] "Generic (PLEG): container finished" podID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerID="f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a" exitCode=0 Dec 02 10:41:51 crc kubenswrapper[4679]: I1202 10:41:51.148133 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerDied","Data":"f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a"} Dec 02 10:41:51 crc kubenswrapper[4679]: I1202 10:41:51.148410 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerStarted","Data":"bd536c448f687f0f54484473ed7a8f220c6610cb316b6c050a41bba568cbc5ac"} Dec 02 10:41:53 crc kubenswrapper[4679]: I1202 10:41:53.164532 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerStarted","Data":"6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462"} Dec 02 10:41:54 crc kubenswrapper[4679]: I1202 10:41:54.177666 4679 generic.go:334] "Generic (PLEG): container finished" podID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerID="6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462" exitCode=0 Dec 02 10:41:54 crc kubenswrapper[4679]: I1202 10:41:54.177712 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerDied","Data":"6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462"} Dec 02 10:41:55 crc kubenswrapper[4679]: I1202 10:41:55.502854 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 10:41:55 crc kubenswrapper[4679]: I1202 10:41:55.503489 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 10:41:55 crc kubenswrapper[4679]: I1202 10:41:55.521348 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 10:41:55 crc kubenswrapper[4679]: I1202 10:41:55.557611 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 10:41:56 crc kubenswrapper[4679]: I1202 10:41:56.244693 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 10:41:56 crc kubenswrapper[4679]: I1202 10:41:56.517462 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d049f08-8029-4ef7-9383-34925ba3d48b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:56 crc kubenswrapper[4679]: I1202 10:41:56.517705 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d049f08-8029-4ef7-9383-34925ba3d48b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:57 crc kubenswrapper[4679]: I1202 10:41:57.214811 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerStarted","Data":"0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858"} Dec 02 10:41:57 crc kubenswrapper[4679]: I1202 10:41:57.238767 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdmfv" podStartSLOduration=3.398824194 podStartE2EDuration="8.238751365s" podCreationTimestamp="2025-12-02 10:41:49 +0000 UTC" firstStartedPulling="2025-12-02 10:41:51.14985687 +0000 UTC m=+1364.479995730" lastFinishedPulling="2025-12-02 10:41:55.989784041 +0000 UTC m=+1369.319922901" observedRunningTime="2025-12-02 10:41:57.235144191 +0000 UTC m=+1370.565283051" watchObservedRunningTime="2025-12-02 10:41:57.238751365 +0000 UTC m=+1370.568890225" Dec 02 10:41:57 crc kubenswrapper[4679]: I1202 10:41:57.524861 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:41:57 crc kubenswrapper[4679]: I1202 10:41:57.524901 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 10:41:58 crc kubenswrapper[4679]: I1202 10:41:58.539637 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81f22355-2ad7-4fd6-bc88-779cd704b038" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:58 crc kubenswrapper[4679]: I1202 10:41:58.540069 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81f22355-2ad7-4fd6-bc88-779cd704b038" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 10:41:59 crc kubenswrapper[4679]: I1202 10:41:59.838870 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:41:59 crc kubenswrapper[4679]: I1202 10:41:59.839840 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:42:00 crc kubenswrapper[4679]: I1202 10:42:00.937112 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bdmfv" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="registry-server" probeResult="failure" output=< Dec 02 10:42:00 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 10:42:00 crc kubenswrapper[4679]: > Dec 02 10:42:05 crc kubenswrapper[4679]: I1202 10:42:05.511345 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 10:42:05 crc kubenswrapper[4679]: I1202 10:42:05.515448 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 10:42:05 crc kubenswrapper[4679]: I1202 10:42:05.518673 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 10:42:06 crc kubenswrapper[4679]: I1202 10:42:06.322156 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 10:42:06 crc kubenswrapper[4679]: I1202 10:42:06.399750 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 02 10:42:07 crc kubenswrapper[4679]: I1202 10:42:07.530105 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 10:42:07 crc kubenswrapper[4679]: I1202 10:42:07.530936 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 10:42:07 crc kubenswrapper[4679]: I1202 10:42:07.537590 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 10:42:07 crc kubenswrapper[4679]: I1202 10:42:07.542692 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 10:42:08 crc kubenswrapper[4679]: I1202 10:42:08.335241 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 10:42:08 crc kubenswrapper[4679]: I1202 10:42:08.345851 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 10:42:09 crc kubenswrapper[4679]: I1202 10:42:09.899406 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:42:09 crc kubenswrapper[4679]: I1202 10:42:09.960548 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:42:10 crc kubenswrapper[4679]: I1202 10:42:10.139003 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdmfv"] Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.361290 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bdmfv" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="registry-server" containerID="cri-o://0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858" gracePeriod=2 Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.843649 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.982736 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw6w8\" (UniqueName: \"kubernetes.io/projected/97c9b812-8677-4b11-96f5-b3537a7b9ab2-kube-api-access-lw6w8\") pod \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.982914 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-utilities\") pod \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.982964 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-catalog-content\") pod \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\" (UID: \"97c9b812-8677-4b11-96f5-b3537a7b9ab2\") " Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.983654 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-utilities" (OuterVolumeSpecName: "utilities") pod "97c9b812-8677-4b11-96f5-b3537a7b9ab2" (UID: "97c9b812-8677-4b11-96f5-b3537a7b9ab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.983926 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:11 crc kubenswrapper[4679]: I1202 10:42:11.996151 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c9b812-8677-4b11-96f5-b3537a7b9ab2-kube-api-access-lw6w8" (OuterVolumeSpecName: "kube-api-access-lw6w8") pod "97c9b812-8677-4b11-96f5-b3537a7b9ab2" (UID: "97c9b812-8677-4b11-96f5-b3537a7b9ab2"). InnerVolumeSpecName "kube-api-access-lw6w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.081531 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97c9b812-8677-4b11-96f5-b3537a7b9ab2" (UID: "97c9b812-8677-4b11-96f5-b3537a7b9ab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.085574 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw6w8\" (UniqueName: \"kubernetes.io/projected/97c9b812-8677-4b11-96f5-b3537a7b9ab2-kube-api-access-lw6w8\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.085616 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c9b812-8677-4b11-96f5-b3537a7b9ab2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.371960 4679 generic.go:334] "Generic (PLEG): container finished" podID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerID="0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858" exitCode=0 Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.372005 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerDied","Data":"0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858"} Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.372031 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdmfv" event={"ID":"97c9b812-8677-4b11-96f5-b3537a7b9ab2","Type":"ContainerDied","Data":"bd536c448f687f0f54484473ed7a8f220c6610cb316b6c050a41bba568cbc5ac"} Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.372047 4679 scope.go:117] "RemoveContainer" containerID="0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.372182 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdmfv" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.413764 4679 scope.go:117] "RemoveContainer" containerID="6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.414265 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdmfv"] Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.422385 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bdmfv"] Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.439917 4679 scope.go:117] "RemoveContainer" containerID="f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.477687 4679 scope.go:117] "RemoveContainer" containerID="0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858" Dec 02 10:42:12 crc kubenswrapper[4679]: E1202 10:42:12.478206 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858\": container with ID starting with 0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858 not found: ID does not exist" containerID="0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.478238 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858"} err="failed to get container status \"0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858\": rpc error: code = NotFound desc = could not find container \"0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858\": container with ID starting with 0343e62a08564b0684a1a91921d5085fc46726df548189561b232a40b5190858 not found: ID does not exist" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.478259 4679 scope.go:117] "RemoveContainer" containerID="6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462" Dec 02 10:42:12 crc kubenswrapper[4679]: E1202 10:42:12.478777 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462\": container with ID starting with 6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462 not found: ID does not exist" containerID="6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.478801 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462"} err="failed to get container status \"6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462\": rpc error: code = NotFound desc = could not find container \"6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462\": container with ID starting with 6319eeaea67758ffa1853f81bedddde6cb78e8c9e6996709a59068f08ff97462 not found: ID does not exist" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.478815 4679 scope.go:117] "RemoveContainer" containerID="f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a" Dec 02 10:42:12 crc kubenswrapper[4679]: E1202 10:42:12.479075 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a\": container with ID starting with f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a not found: ID does not exist" containerID="f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.479094 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a"} err="failed to get container status \"f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a\": rpc error: code = NotFound desc = could not find container \"f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a\": container with ID starting with f89d58fc2952abae321cb9b6155b4c4179b3420d6f43fba7f758e88a21a38a4a not found: ID does not exist" Dec 02 10:42:12 crc kubenswrapper[4679]: I1202 10:42:12.923742 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" path="/var/lib/kubelet/pods/97c9b812-8677-4b11-96f5-b3537a7b9ab2/volumes" Dec 02 10:42:16 crc kubenswrapper[4679]: I1202 10:42:16.495041 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:42:17 crc kubenswrapper[4679]: I1202 10:42:17.430335 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:42:20 crc kubenswrapper[4679]: I1202 10:42:20.520953 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="rabbitmq" containerID="cri-o://b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb" gracePeriod=604796 Dec 02 10:42:21 crc kubenswrapper[4679]: I1202 10:42:21.182001 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="rabbitmq" containerID="cri-o://0258d496799a824dccc899444561afc95538f8181fcb3bbbe5c69aa6d0d1d35c" gracePeriod=604797 Dec 02 10:42:25 crc kubenswrapper[4679]: I1202 10:42:25.512411 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 02 10:42:25 crc kubenswrapper[4679]: I1202 10:42:25.987523 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.085141 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.164818 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w6pp\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-kube-api-access-8w6pp\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165091 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-erlang-cookie\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165128 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165158 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-erlang-cookie-secret\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165202 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-confd\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165320 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-config-data\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165352 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-pod-info\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165376 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-plugins-conf\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165407 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-tls\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165516 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-server-conf\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165572 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-plugins\") pod \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\" (UID: \"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae\") " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.165665 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.166081 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.166429 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.166746 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.172817 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-kube-api-access-8w6pp" (OuterVolumeSpecName: "kube-api-access-8w6pp") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "kube-api-access-8w6pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.174846 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.180424 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.186677 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-pod-info" (OuterVolumeSpecName: "pod-info") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.186832 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.224870 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-config-data" (OuterVolumeSpecName: "config-data") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.260877 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-server-conf" (OuterVolumeSpecName: "server-conf") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268129 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268163 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w6pp\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-kube-api-access-8w6pp\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268193 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268202 4679 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268210 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268220 4679 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-pod-info\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268228 4679 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268237 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.268247 4679 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-server-conf\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.294995 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.339511 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" (UID: "66a56dd2-b866-4b5b-9fd5-94e3dc8422ae"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.370488 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.370520 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.515836 4679 generic.go:334] "Generic (PLEG): container finished" podID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerID="0258d496799a824dccc899444561afc95538f8181fcb3bbbe5c69aa6d0d1d35c" exitCode=0 Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.515921 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"416d04f1-7afb-4bdf-a5d5-1a0c381513f9","Type":"ContainerDied","Data":"0258d496799a824dccc899444561afc95538f8181fcb3bbbe5c69aa6d0d1d35c"} Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.517916 4679 generic.go:334] "Generic (PLEG): container finished" podID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerID="b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb" exitCode=0 Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.517948 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae","Type":"ContainerDied","Data":"b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb"} Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.517964 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"66a56dd2-b866-4b5b-9fd5-94e3dc8422ae","Type":"ContainerDied","Data":"01c828f2474a6ee870a17d1e49a3ca2a84c45a85740075857e43ac83bb0cc680"} Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.517980 4679 scope.go:117] "RemoveContainer" containerID="b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.517984 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.543856 4679 scope.go:117] "RemoveContainer" containerID="b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.562396 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.581864 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.602934 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.603473 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="setup-container" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.603509 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="setup-container" Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.603522 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="registry-server" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.603530 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="registry-server" Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.603548 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="rabbitmq" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.603558 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="rabbitmq" Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.603569 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="extract-content" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.603576 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="extract-content" Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.603599 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="extract-utilities" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.603606 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="extract-utilities" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.608947 4679 scope.go:117] "RemoveContainer" containerID="b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb" Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.612052 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb\": container with ID starting with b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb not found: ID does not exist" containerID="b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.612099 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb"} err="failed to get container status \"b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb\": rpc error: code = NotFound desc = could not find container \"b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb\": container with ID starting with b13b18bf1ed8738c2bddf52e67eca83d4c36548e68f2ea724c6e92668513a5bb not found: ID does not exist" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.612127 4679 scope.go:117] "RemoveContainer" containerID="b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f" Dec 02 10:42:27 crc kubenswrapper[4679]: E1202 10:42:27.613550 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f\": container with ID starting with b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f not found: ID does not exist" containerID="b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.613606 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f"} err="failed to get container status \"b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f\": rpc error: code = NotFound desc = could not find container \"b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f\": container with ID starting with b7aaf87ac95fc203ff8cc459fd2bcee962d0d79592033c2dbc8ef6125aed046f not found: ID does not exist" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.617711 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" containerName="rabbitmq" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.617739 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c9b812-8677-4b11-96f5-b3537a7b9ab2" containerName="registry-server" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.624590 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.624702 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.627042 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.627485 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.627631 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.627829 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.627849 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.627977 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.628018 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2dqtv" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778034 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778081 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778107 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-config-data\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778138 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-server-conf\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778154 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778177 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/561abb24-2326-4992-b863-6065fbdcb959-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778255 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktw9\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-kube-api-access-fktw9\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778284 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778322 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/561abb24-2326-4992-b863-6065fbdcb959-pod-info\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778338 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.778356 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879781 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktw9\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-kube-api-access-fktw9\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879833 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879860 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/561abb24-2326-4992-b863-6065fbdcb959-pod-info\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879878 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879895 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879942 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879964 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.879987 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-config-data\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.880017 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-server-conf\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.880031 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.880052 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/561abb24-2326-4992-b863-6065fbdcb959-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.880634 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.881333 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.881493 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.881524 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-config-data\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.881735 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.882173 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/561abb24-2326-4992-b863-6065fbdcb959-server-conf\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.886586 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/561abb24-2326-4992-b863-6065fbdcb959-pod-info\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.886595 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.886942 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/561abb24-2326-4992-b863-6065fbdcb959-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.887251 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.897445 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktw9\" (UniqueName: \"kubernetes.io/projected/561abb24-2326-4992-b863-6065fbdcb959-kube-api-access-fktw9\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.928073 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"561abb24-2326-4992-b863-6065fbdcb959\") " pod="openstack/rabbitmq-server-0" Dec 02 10:42:27 crc kubenswrapper[4679]: I1202 10:42:27.956835 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.023099 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.082808 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6nl2\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-kube-api-access-d6nl2\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.082862 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-tls\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.082900 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-plugins-conf\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.082935 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-config-data\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.082959 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-erlang-cookie\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.082984 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-server-conf\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.083004 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.083039 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-confd\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.083058 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-pod-info\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.083075 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-plugins\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.083103 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-erlang-cookie-secret\") pod \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\" (UID: \"416d04f1-7afb-4bdf-a5d5-1a0c381513f9\") " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.084180 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.086387 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.086766 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.088958 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.090629 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-kube-api-access-d6nl2" (OuterVolumeSpecName: "kube-api-access-d6nl2") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "kube-api-access-d6nl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.090730 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.096398 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.097601 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-pod-info" (OuterVolumeSpecName: "pod-info") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.162799 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-config-data" (OuterVolumeSpecName: "config-data") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.169908 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-server-conf" (OuterVolumeSpecName: "server-conf") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185406 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6nl2\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-kube-api-access-d6nl2\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185446 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185460 4679 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185471 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185485 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185498 4679 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185525 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185537 4679 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185548 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.185559 4679 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.228225 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.269714 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "416d04f1-7afb-4bdf-a5d5-1a0c381513f9" (UID: "416d04f1-7afb-4bdf-a5d5-1a0c381513f9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.288141 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.288188 4679 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/416d04f1-7afb-4bdf-a5d5-1a0c381513f9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.512084 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 10:42:28 crc kubenswrapper[4679]: W1202 10:42:28.519545 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod561abb24_2326_4992_b863_6065fbdcb959.slice/crio-ebb9acef7148e3ff5a16b73b96e69fbde657988bf7678aec384e84d1ba93a4ed WatchSource:0}: Error finding container ebb9acef7148e3ff5a16b73b96e69fbde657988bf7678aec384e84d1ba93a4ed: Status 404 returned error can't find the container with id ebb9acef7148e3ff5a16b73b96e69fbde657988bf7678aec384e84d1ba93a4ed Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.536681 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"416d04f1-7afb-4bdf-a5d5-1a0c381513f9","Type":"ContainerDied","Data":"df03b283b858a9fd75c51bacb1b6f3af4a33b66d4389397c7dd1422fed752005"} Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.536765 4679 scope.go:117] "RemoveContainer" containerID="0258d496799a824dccc899444561afc95538f8181fcb3bbbe5c69aa6d0d1d35c" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.536703 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.575386 4679 scope.go:117] "RemoveContainer" containerID="df9a6c6da390576c9b707919a8e3e5b6cf8c75ee5cff628082a007e62d19b02d" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.584966 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.593339 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.617408 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:42:28 crc kubenswrapper[4679]: E1202 10:42:28.617877 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="rabbitmq" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.617899 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="rabbitmq" Dec 02 10:42:28 crc kubenswrapper[4679]: E1202 10:42:28.617939 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="setup-container" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.617946 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="setup-container" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.618166 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" containerName="rabbitmq" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.619528 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.625047 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.625245 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.625445 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.625529 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-78jjf" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.625645 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.625724 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.626336 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.640607 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695690 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695777 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/17e32beb-4297-4540-a48b-b1d3d4a32ee2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695800 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695818 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr8q8\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-kube-api-access-tr8q8\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695843 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/17e32beb-4297-4540-a48b-b1d3d4a32ee2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695865 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695944 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695968 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.695997 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.696015 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.696035 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797675 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797739 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797764 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797788 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797842 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797908 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/17e32beb-4297-4540-a48b-b1d3d4a32ee2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797928 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797946 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr8q8\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-kube-api-access-tr8q8\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797966 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/17e32beb-4297-4540-a48b-b1d3d4a32ee2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.797986 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.798028 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.798492 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.798960 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.799219 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.799259 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.799497 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.799999 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17e32beb-4297-4540-a48b-b1d3d4a32ee2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.802212 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.802358 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/17e32beb-4297-4540-a48b-b1d3d4a32ee2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.802531 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.803895 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/17e32beb-4297-4540-a48b-b1d3d4a32ee2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.818010 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr8q8\" (UniqueName: \"kubernetes.io/projected/17e32beb-4297-4540-a48b-b1d3d4a32ee2-kube-api-access-tr8q8\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.830637 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17e32beb-4297-4540-a48b-b1d3d4a32ee2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.920673 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="416d04f1-7afb-4bdf-a5d5-1a0c381513f9" path="/var/lib/kubelet/pods/416d04f1-7afb-4bdf-a5d5-1a0c381513f9/volumes" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.922044 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66a56dd2-b866-4b5b-9fd5-94e3dc8422ae" path="/var/lib/kubelet/pods/66a56dd2-b866-4b5b-9fd5-94e3dc8422ae/volumes" Dec 02 10:42:28 crc kubenswrapper[4679]: I1202 10:42:28.955877 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:42:29 crc kubenswrapper[4679]: I1202 10:42:29.456124 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 10:42:29 crc kubenswrapper[4679]: W1202 10:42:29.459359 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17e32beb_4297_4540_a48b_b1d3d4a32ee2.slice/crio-8e12db7d9452857fae0cb3156d94d0aafb027be52e847f8b6ab293ff9dadafcb WatchSource:0}: Error finding container 8e12db7d9452857fae0cb3156d94d0aafb027be52e847f8b6ab293ff9dadafcb: Status 404 returned error can't find the container with id 8e12db7d9452857fae0cb3156d94d0aafb027be52e847f8b6ab293ff9dadafcb Dec 02 10:42:29 crc kubenswrapper[4679]: I1202 10:42:29.553036 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"561abb24-2326-4992-b863-6065fbdcb959","Type":"ContainerStarted","Data":"ebb9acef7148e3ff5a16b73b96e69fbde657988bf7678aec384e84d1ba93a4ed"} Dec 02 10:42:29 crc kubenswrapper[4679]: I1202 10:42:29.554971 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17e32beb-4297-4540-a48b-b1d3d4a32ee2","Type":"ContainerStarted","Data":"8e12db7d9452857fae0cb3156d94d0aafb027be52e847f8b6ab293ff9dadafcb"} Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.327414 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cg2r9"] Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.336726 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.339953 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.362158 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cg2r9"] Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.426924 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.426986 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.427008 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.427052 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.427066 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snjtk\" (UniqueName: \"kubernetes.io/projected/a73d80ad-6be2-4be1-a16e-d992825a5911-kube-api-access-snjtk\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.427091 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-config\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.427188 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-svc\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528737 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528810 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528833 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528852 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528868 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snjtk\" (UniqueName: \"kubernetes.io/projected/a73d80ad-6be2-4be1-a16e-d992825a5911-kube-api-access-snjtk\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528899 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-config\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.528954 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-svc\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.529993 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.530025 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.529998 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.530074 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-svc\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.530560 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-config\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.531034 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.548467 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snjtk\" (UniqueName: \"kubernetes.io/projected/a73d80ad-6be2-4be1-a16e-d992825a5911-kube-api-access-snjtk\") pod \"dnsmasq-dns-67b789f86c-cg2r9\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.567574 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"561abb24-2326-4992-b863-6065fbdcb959","Type":"ContainerStarted","Data":"756d703a7f67d15504b1d31d3743f16d2b2fd1778dad7b6e09d7ddf5efaca80e"} Dec 02 10:42:30 crc kubenswrapper[4679]: I1202 10:42:30.653923 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:31 crc kubenswrapper[4679]: I1202 10:42:31.230862 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cg2r9"] Dec 02 10:42:31 crc kubenswrapper[4679]: W1202 10:42:31.233382 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda73d80ad_6be2_4be1_a16e_d992825a5911.slice/crio-396442d4dc55e4d5df24677afd84a4fcbf5316cc74703e0a9c94250bf937efc7 WatchSource:0}: Error finding container 396442d4dc55e4d5df24677afd84a4fcbf5316cc74703e0a9c94250bf937efc7: Status 404 returned error can't find the container with id 396442d4dc55e4d5df24677afd84a4fcbf5316cc74703e0a9c94250bf937efc7 Dec 02 10:42:31 crc kubenswrapper[4679]: I1202 10:42:31.584656 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" event={"ID":"a73d80ad-6be2-4be1-a16e-d992825a5911","Type":"ContainerStarted","Data":"396442d4dc55e4d5df24677afd84a4fcbf5316cc74703e0a9c94250bf937efc7"} Dec 02 10:42:31 crc kubenswrapper[4679]: I1202 10:42:31.587456 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17e32beb-4297-4540-a48b-b1d3d4a32ee2","Type":"ContainerStarted","Data":"366323a947f13374a73843b363a2f548bd3998db448120b75de793071f48213d"} Dec 02 10:42:32 crc kubenswrapper[4679]: I1202 10:42:32.598338 4679 generic.go:334] "Generic (PLEG): container finished" podID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerID="fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5" exitCode=0 Dec 02 10:42:32 crc kubenswrapper[4679]: I1202 10:42:32.598418 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" event={"ID":"a73d80ad-6be2-4be1-a16e-d992825a5911","Type":"ContainerDied","Data":"fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5"} Dec 02 10:42:33 crc kubenswrapper[4679]: I1202 10:42:33.609391 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" event={"ID":"a73d80ad-6be2-4be1-a16e-d992825a5911","Type":"ContainerStarted","Data":"c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a"} Dec 02 10:42:33 crc kubenswrapper[4679]: I1202 10:42:33.609723 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:33 crc kubenswrapper[4679]: I1202 10:42:33.651701 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" podStartSLOduration=3.651669596 podStartE2EDuration="3.651669596s" podCreationTimestamp="2025-12-02 10:42:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:42:33.644639064 +0000 UTC m=+1406.974777924" watchObservedRunningTime="2025-12-02 10:42:33.651669596 +0000 UTC m=+1406.981808496" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.655606 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.735844 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-kj97s"] Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.736116 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerName="dnsmasq-dns" containerID="cri-o://b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a" gracePeriod=10 Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.850348 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-b4d9r"] Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.855749 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.886932 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-b4d9r"] Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957108 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkt62\" (UniqueName: \"kubernetes.io/projected/c203b3f2-0c27-4421-9912-9ac96318189c-kube-api-access-gkt62\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957166 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957259 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957338 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957366 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-config\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957435 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:40 crc kubenswrapper[4679]: I1202 10:42:40.957462 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.058915 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.058981 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-config\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.059040 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.059074 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.059156 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkt62\" (UniqueName: \"kubernetes.io/projected/c203b3f2-0c27-4421-9912-9ac96318189c-kube-api-access-gkt62\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.059185 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.059288 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.060330 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-config\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.061624 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.061725 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.062112 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.062317 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.062365 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c203b3f2-0c27-4421-9912-9ac96318189c-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.080921 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkt62\" (UniqueName: \"kubernetes.io/projected/c203b3f2-0c27-4421-9912-9ac96318189c-kube-api-access-gkt62\") pod \"dnsmasq-dns-cb6ffcf87-b4d9r\" (UID: \"c203b3f2-0c27-4421-9912-9ac96318189c\") " pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.218414 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.228645 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.365262 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-config\") pod \"03f46a70-31b1-49ee-ab7a-ebd520592409\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.365486 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4dj8\" (UniqueName: \"kubernetes.io/projected/03f46a70-31b1-49ee-ab7a-ebd520592409-kube-api-access-l4dj8\") pod \"03f46a70-31b1-49ee-ab7a-ebd520592409\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.365583 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-svc\") pod \"03f46a70-31b1-49ee-ab7a-ebd520592409\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.365624 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-nb\") pod \"03f46a70-31b1-49ee-ab7a-ebd520592409\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.365656 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-swift-storage-0\") pod \"03f46a70-31b1-49ee-ab7a-ebd520592409\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.365722 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-sb\") pod \"03f46a70-31b1-49ee-ab7a-ebd520592409\" (UID: \"03f46a70-31b1-49ee-ab7a-ebd520592409\") " Dec 02 10:42:41 crc kubenswrapper[4679]: I1202 10:42:41.378773 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03f46a70-31b1-49ee-ab7a-ebd520592409-kube-api-access-l4dj8" (OuterVolumeSpecName: "kube-api-access-l4dj8") pod "03f46a70-31b1-49ee-ab7a-ebd520592409" (UID: "03f46a70-31b1-49ee-ab7a-ebd520592409"). InnerVolumeSpecName "kube-api-access-l4dj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.418769 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03f46a70-31b1-49ee-ab7a-ebd520592409" (UID: "03f46a70-31b1-49ee-ab7a-ebd520592409"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.424885 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-config" (OuterVolumeSpecName: "config") pod "03f46a70-31b1-49ee-ab7a-ebd520592409" (UID: "03f46a70-31b1-49ee-ab7a-ebd520592409"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.425758 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "03f46a70-31b1-49ee-ab7a-ebd520592409" (UID: "03f46a70-31b1-49ee-ab7a-ebd520592409"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.443555 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "03f46a70-31b1-49ee-ab7a-ebd520592409" (UID: "03f46a70-31b1-49ee-ab7a-ebd520592409"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.453128 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "03f46a70-31b1-49ee-ab7a-ebd520592409" (UID: "03f46a70-31b1-49ee-ab7a-ebd520592409"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.467700 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.467744 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.467756 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.467765 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.467773 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03f46a70-31b1-49ee-ab7a-ebd520592409-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.467783 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4dj8\" (UniqueName: \"kubernetes.io/projected/03f46a70-31b1-49ee-ab7a-ebd520592409-kube-api-access-l4dj8\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.704319 4679 generic.go:334] "Generic (PLEG): container finished" podID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerID="b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a" exitCode=0 Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.704336 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" event={"ID":"03f46a70-31b1-49ee-ab7a-ebd520592409","Type":"ContainerDied","Data":"b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a"} Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.704382 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" event={"ID":"03f46a70-31b1-49ee-ab7a-ebd520592409","Type":"ContainerDied","Data":"eeb85bfba02139b8245af19e5a8a7532cd217ecd23fb63acf596f22c02fa6bf5"} Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.704403 4679 scope.go:117] "RemoveContainer" containerID="b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.704347 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-kj97s" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.731364 4679 scope.go:117] "RemoveContainer" containerID="a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.738225 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-kj97s"] Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.748103 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-kj97s"] Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.789015 4679 scope.go:117] "RemoveContainer" containerID="b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a" Dec 02 10:42:42 crc kubenswrapper[4679]: E1202 10:42:41.789576 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a\": container with ID starting with b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a not found: ID does not exist" containerID="b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.789602 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a"} err="failed to get container status \"b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a\": rpc error: code = NotFound desc = could not find container \"b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a\": container with ID starting with b3b1c59c6ce8d9b8e0a011871d97c5101da925efbe389783d341835495ecc29a not found: ID does not exist" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.789622 4679 scope.go:117] "RemoveContainer" containerID="a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5" Dec 02 10:42:42 crc kubenswrapper[4679]: E1202 10:42:41.790001 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5\": container with ID starting with a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5 not found: ID does not exist" containerID="a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:41.790048 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5"} err="failed to get container status \"a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5\": rpc error: code = NotFound desc = could not find container \"a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5\": container with ID starting with a3162029ac75b19771d925292a2c607361364ea39f08ec828ec0cf1deaacd8a5 not found: ID does not exist" Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:42.505143 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-b4d9r"] Dec 02 10:42:42 crc kubenswrapper[4679]: W1202 10:42:42.507024 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc203b3f2_0c27_4421_9912_9ac96318189c.slice/crio-bce7cfb7f25192a6c5c358d896c3284b2f43bd0a6b2d64b8f3dc003ff979ab78 WatchSource:0}: Error finding container bce7cfb7f25192a6c5c358d896c3284b2f43bd0a6b2d64b8f3dc003ff979ab78: Status 404 returned error can't find the container with id bce7cfb7f25192a6c5c358d896c3284b2f43bd0a6b2d64b8f3dc003ff979ab78 Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:42.714718 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" event={"ID":"c203b3f2-0c27-4421-9912-9ac96318189c","Type":"ContainerStarted","Data":"bce7cfb7f25192a6c5c358d896c3284b2f43bd0a6b2d64b8f3dc003ff979ab78"} Dec 02 10:42:42 crc kubenswrapper[4679]: I1202 10:42:42.930083 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" path="/var/lib/kubelet/pods/03f46a70-31b1-49ee-ab7a-ebd520592409/volumes" Dec 02 10:42:43 crc kubenswrapper[4679]: I1202 10:42:43.728087 4679 generic.go:334] "Generic (PLEG): container finished" podID="c203b3f2-0c27-4421-9912-9ac96318189c" containerID="fee33e173564a2313bc4eebe3b07d5950764cfaabd4fab3615580caad357a9a1" exitCode=0 Dec 02 10:42:43 crc kubenswrapper[4679]: I1202 10:42:43.728155 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" event={"ID":"c203b3f2-0c27-4421-9912-9ac96318189c","Type":"ContainerDied","Data":"fee33e173564a2313bc4eebe3b07d5950764cfaabd4fab3615580caad357a9a1"} Dec 02 10:42:44 crc kubenswrapper[4679]: I1202 10:42:44.738614 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" event={"ID":"c203b3f2-0c27-4421-9912-9ac96318189c","Type":"ContainerStarted","Data":"7fc1419d97b2bfa856306304190bcb937f831c4ed481c079dfc425eb0a77082d"} Dec 02 10:42:44 crc kubenswrapper[4679]: I1202 10:42:44.738975 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:44 crc kubenswrapper[4679]: I1202 10:42:44.764486 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" podStartSLOduration=4.764467552 podStartE2EDuration="4.764467552s" podCreationTimestamp="2025-12-02 10:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:42:44.75884018 +0000 UTC m=+1418.088979040" watchObservedRunningTime="2025-12-02 10:42:44.764467552 +0000 UTC m=+1418.094606412" Dec 02 10:42:46 crc kubenswrapper[4679]: I1202 10:42:46.931420 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:42:46 crc kubenswrapper[4679]: I1202 10:42:46.931786 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.220692 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-b4d9r" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.297065 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cg2r9"] Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.297328 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerName="dnsmasq-dns" containerID="cri-o://c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a" gracePeriod=10 Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.805925 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.812409 4679 generic.go:334] "Generic (PLEG): container finished" podID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerID="c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a" exitCode=0 Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.812464 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" event={"ID":"a73d80ad-6be2-4be1-a16e-d992825a5911","Type":"ContainerDied","Data":"c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a"} Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.812497 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" event={"ID":"a73d80ad-6be2-4be1-a16e-d992825a5911","Type":"ContainerDied","Data":"396442d4dc55e4d5df24677afd84a4fcbf5316cc74703e0a9c94250bf937efc7"} Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.812521 4679 scope.go:117] "RemoveContainer" containerID="c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.812703 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cg2r9" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.835573 4679 scope.go:117] "RemoveContainer" containerID="fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.858748 4679 scope.go:117] "RemoveContainer" containerID="c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a" Dec 02 10:42:51 crc kubenswrapper[4679]: E1202 10:42:51.859915 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a\": container with ID starting with c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a not found: ID does not exist" containerID="c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.859974 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a"} err="failed to get container status \"c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a\": rpc error: code = NotFound desc = could not find container \"c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a\": container with ID starting with c85764ea9de9f607f53bd8b44442b5d990830ffa6968fb26c777e0b72190401a not found: ID does not exist" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.859999 4679 scope.go:117] "RemoveContainer" containerID="fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5" Dec 02 10:42:51 crc kubenswrapper[4679]: E1202 10:42:51.860508 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5\": container with ID starting with fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5 not found: ID does not exist" containerID="fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.860557 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5"} err="failed to get container status \"fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5\": rpc error: code = NotFound desc = could not find container \"fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5\": container with ID starting with fec5c29957eef8bc8010a3eb14f2d54cd56f42e830901575066274aa724f85a5 not found: ID does not exist" Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.980874 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-openstack-edpm-ipam\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.981419 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-config\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.981456 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snjtk\" (UniqueName: \"kubernetes.io/projected/a73d80ad-6be2-4be1-a16e-d992825a5911-kube-api-access-snjtk\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.981485 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-nb\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.981531 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-svc\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.981573 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-sb\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.981590 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-swift-storage-0\") pod \"a73d80ad-6be2-4be1-a16e-d992825a5911\" (UID: \"a73d80ad-6be2-4be1-a16e-d992825a5911\") " Dec 02 10:42:51 crc kubenswrapper[4679]: I1202 10:42:51.989907 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73d80ad-6be2-4be1-a16e-d992825a5911-kube-api-access-snjtk" (OuterVolumeSpecName: "kube-api-access-snjtk") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "kube-api-access-snjtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.031073 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.032475 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.037666 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.041855 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.047823 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-config" (OuterVolumeSpecName: "config") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.058454 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a73d80ad-6be2-4be1-a16e-d992825a5911" (UID: "a73d80ad-6be2-4be1-a16e-d992825a5911"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087839 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087896 4679 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-config\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087916 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snjtk\" (UniqueName: \"kubernetes.io/projected/a73d80ad-6be2-4be1-a16e-d992825a5911-kube-api-access-snjtk\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087935 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087952 4679 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087970 4679 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.087986 4679 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a73d80ad-6be2-4be1-a16e-d992825a5911-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.151060 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cg2r9"] Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.158277 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cg2r9"] Dec 02 10:42:52 crc kubenswrapper[4679]: I1202 10:42:52.920016 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" path="/var/lib/kubelet/pods/a73d80ad-6be2-4be1-a16e-d992825a5911/volumes" Dec 02 10:43:02 crc kubenswrapper[4679]: I1202 10:43:02.919883 4679 generic.go:334] "Generic (PLEG): container finished" podID="561abb24-2326-4992-b863-6065fbdcb959" containerID="756d703a7f67d15504b1d31d3743f16d2b2fd1778dad7b6e09d7ddf5efaca80e" exitCode=0 Dec 02 10:43:02 crc kubenswrapper[4679]: I1202 10:43:02.919942 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"561abb24-2326-4992-b863-6065fbdcb959","Type":"ContainerDied","Data":"756d703a7f67d15504b1d31d3743f16d2b2fd1778dad7b6e09d7ddf5efaca80e"} Dec 02 10:43:03 crc kubenswrapper[4679]: I1202 10:43:03.943675 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"561abb24-2326-4992-b863-6065fbdcb959","Type":"ContainerStarted","Data":"0667752cb33c9e129a5680caca6b543a2805ec98e5ba169b8cea9690901b80e5"} Dec 02 10:43:03 crc kubenswrapper[4679]: I1202 10:43:03.945586 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 02 10:43:03 crc kubenswrapper[4679]: I1202 10:43:03.954101 4679 generic.go:334] "Generic (PLEG): container finished" podID="17e32beb-4297-4540-a48b-b1d3d4a32ee2" containerID="366323a947f13374a73843b363a2f548bd3998db448120b75de793071f48213d" exitCode=0 Dec 02 10:43:03 crc kubenswrapper[4679]: I1202 10:43:03.954211 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17e32beb-4297-4540-a48b-b1d3d4a32ee2","Type":"ContainerDied","Data":"366323a947f13374a73843b363a2f548bd3998db448120b75de793071f48213d"} Dec 02 10:43:03 crc kubenswrapper[4679]: I1202 10:43:03.992845 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.992826439 podStartE2EDuration="36.992826439s" podCreationTimestamp="2025-12-02 10:42:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:43:03.986488776 +0000 UTC m=+1437.316627656" watchObservedRunningTime="2025-12-02 10:43:03.992826439 +0000 UTC m=+1437.322965309" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.546855 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6"] Dec 02 10:43:04 crc kubenswrapper[4679]: E1202 10:43:04.547630 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerName="init" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.547653 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerName="init" Dec 02 10:43:04 crc kubenswrapper[4679]: E1202 10:43:04.547670 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerName="dnsmasq-dns" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.547676 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerName="dnsmasq-dns" Dec 02 10:43:04 crc kubenswrapper[4679]: E1202 10:43:04.547686 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerName="dnsmasq-dns" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.547693 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerName="dnsmasq-dns" Dec 02 10:43:04 crc kubenswrapper[4679]: E1202 10:43:04.547699 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerName="init" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.547705 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerName="init" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.547936 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73d80ad-6be2-4be1-a16e-d992825a5911" containerName="dnsmasq-dns" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.547965 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="03f46a70-31b1-49ee-ab7a-ebd520592409" containerName="dnsmasq-dns" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.549451 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.553962 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.554245 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.557860 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.558027 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.559887 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6"] Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.634683 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv65f\" (UniqueName: \"kubernetes.io/projected/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-kube-api-access-zv65f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.634766 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.634800 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.634962 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.738200 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.738430 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv65f\" (UniqueName: \"kubernetes.io/projected/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-kube-api-access-zv65f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.738972 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.739422 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.744479 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.744574 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.745727 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.756527 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv65f\" (UniqueName: \"kubernetes.io/projected/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-kube-api-access-zv65f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.864646 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.972419 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17e32beb-4297-4540-a48b-b1d3d4a32ee2","Type":"ContainerStarted","Data":"97938893926f1c3517cbbcb578e04f5bc6388863fdfa3ca7857eef1a4ac54e87"} Dec 02 10:43:04 crc kubenswrapper[4679]: I1202 10:43:04.973290 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:43:05 crc kubenswrapper[4679]: I1202 10:43:05.016130 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.016107828 podStartE2EDuration="37.016107828s" podCreationTimestamp="2025-12-02 10:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 10:43:05.011939238 +0000 UTC m=+1438.342078098" watchObservedRunningTime="2025-12-02 10:43:05.016107828 +0000 UTC m=+1438.346246688" Dec 02 10:43:05 crc kubenswrapper[4679]: I1202 10:43:05.419695 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6"] Dec 02 10:43:05 crc kubenswrapper[4679]: W1202 10:43:05.420509 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52ca33f5_24c7_4dfb_ad4f_6e2b67c1849a.slice/crio-e2f6dd2c67019a839ffd2370cf8102fca2d427d9fd51043b045a39c4b84b96b5 WatchSource:0}: Error finding container e2f6dd2c67019a839ffd2370cf8102fca2d427d9fd51043b045a39c4b84b96b5: Status 404 returned error can't find the container with id e2f6dd2c67019a839ffd2370cf8102fca2d427d9fd51043b045a39c4b84b96b5 Dec 02 10:43:05 crc kubenswrapper[4679]: I1202 10:43:05.985488 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" event={"ID":"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a","Type":"ContainerStarted","Data":"e2f6dd2c67019a839ffd2370cf8102fca2d427d9fd51043b045a39c4b84b96b5"} Dec 02 10:43:15 crc kubenswrapper[4679]: I1202 10:43:15.890596 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:43:16 crc kubenswrapper[4679]: I1202 10:43:16.931676 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:43:16 crc kubenswrapper[4679]: I1202 10:43:16.932103 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:43:17 crc kubenswrapper[4679]: I1202 10:43:17.098800 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" event={"ID":"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a","Type":"ContainerStarted","Data":"80b930c5218f3434ba404ba1fea36b544611d315b5ced652d1468afeae59cc72"} Dec 02 10:43:17 crc kubenswrapper[4679]: I1202 10:43:17.120485 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" podStartSLOduration=2.657409378 podStartE2EDuration="13.120468544s" podCreationTimestamp="2025-12-02 10:43:04 +0000 UTC" firstStartedPulling="2025-12-02 10:43:05.423719765 +0000 UTC m=+1438.753858625" lastFinishedPulling="2025-12-02 10:43:15.886778931 +0000 UTC m=+1449.216917791" observedRunningTime="2025-12-02 10:43:17.11928733 +0000 UTC m=+1450.449426190" watchObservedRunningTime="2025-12-02 10:43:17.120468544 +0000 UTC m=+1450.450607404" Dec 02 10:43:17 crc kubenswrapper[4679]: I1202 10:43:17.959508 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 02 10:43:18 crc kubenswrapper[4679]: I1202 10:43:18.961532 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 02 10:43:20 crc kubenswrapper[4679]: I1202 10:43:20.408248 4679 scope.go:117] "RemoveContainer" containerID="628989f9e7aff7aa3c3ed0a4d7b9f5763065a4db5408848fb320a44d41e906c6" Dec 02 10:43:20 crc kubenswrapper[4679]: I1202 10:43:20.430705 4679 scope.go:117] "RemoveContainer" containerID="e33e703e5278e8dc82542a41a0c09cc2e15e17fddc7874b18591f711340d15ab" Dec 02 10:43:28 crc kubenswrapper[4679]: I1202 10:43:28.193894 4679 generic.go:334] "Generic (PLEG): container finished" podID="52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" containerID="80b930c5218f3434ba404ba1fea36b544611d315b5ced652d1468afeae59cc72" exitCode=0 Dec 02 10:43:28 crc kubenswrapper[4679]: I1202 10:43:28.193974 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" event={"ID":"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a","Type":"ContainerDied","Data":"80b930c5218f3434ba404ba1fea36b544611d315b5ced652d1468afeae59cc72"} Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.624329 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.704898 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-ssh-key\") pod \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.705082 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-inventory\") pod \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.705100 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-repo-setup-combined-ca-bundle\") pod \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.705235 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv65f\" (UniqueName: \"kubernetes.io/projected/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-kube-api-access-zv65f\") pod \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\" (UID: \"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a\") " Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.710983 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" (UID: "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.712813 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-kube-api-access-zv65f" (OuterVolumeSpecName: "kube-api-access-zv65f") pod "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" (UID: "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a"). InnerVolumeSpecName "kube-api-access-zv65f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.732605 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" (UID: "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.738554 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-inventory" (OuterVolumeSpecName: "inventory") pod "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" (UID: "52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.807420 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv65f\" (UniqueName: \"kubernetes.io/projected/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-kube-api-access-zv65f\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.807846 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.807857 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:29 crc kubenswrapper[4679]: I1202 10:43:29.807868 4679 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.212526 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" event={"ID":"52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a","Type":"ContainerDied","Data":"e2f6dd2c67019a839ffd2370cf8102fca2d427d9fd51043b045a39c4b84b96b5"} Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.212581 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.212584 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2f6dd2c67019a839ffd2370cf8102fca2d427d9fd51043b045a39c4b84b96b5" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.295212 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b"] Dec 02 10:43:30 crc kubenswrapper[4679]: E1202 10:43:30.295923 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.295992 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.296256 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.297181 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.300789 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.300926 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.301792 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.302940 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.307447 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b"] Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.418857 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.418958 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.419044 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8klr\" (UniqueName: \"kubernetes.io/projected/e5d79c8f-5612-4597-ba1a-51108bf84927-kube-api-access-s8klr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.520798 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.520885 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.520978 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8klr\" (UniqueName: \"kubernetes.io/projected/e5d79c8f-5612-4597-ba1a-51108bf84927-kube-api-access-s8klr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.526039 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.533133 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.537183 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8klr\" (UniqueName: \"kubernetes.io/projected/e5d79c8f-5612-4597-ba1a-51108bf84927-kube-api-access-s8klr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-bzv4b\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:30 crc kubenswrapper[4679]: I1202 10:43:30.629006 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:31 crc kubenswrapper[4679]: I1202 10:43:31.155633 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b"] Dec 02 10:43:31 crc kubenswrapper[4679]: I1202 10:43:31.221780 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" event={"ID":"e5d79c8f-5612-4597-ba1a-51108bf84927","Type":"ContainerStarted","Data":"023f46fe49531e9828f0d4dbc9cc7d5c4ccbcaaea379f30ea72fb9f6b47bdae0"} Dec 02 10:43:33 crc kubenswrapper[4679]: I1202 10:43:33.243226 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" event={"ID":"e5d79c8f-5612-4597-ba1a-51108bf84927","Type":"ContainerStarted","Data":"4ff9a2c73d0ac0a18f2a3b4dbdcb02c25285adf1b7d715e0efc82afaea777821"} Dec 02 10:43:33 crc kubenswrapper[4679]: I1202 10:43:33.274950 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" podStartSLOduration=2.038155373 podStartE2EDuration="3.274925304s" podCreationTimestamp="2025-12-02 10:43:30 +0000 UTC" firstStartedPulling="2025-12-02 10:43:31.16268894 +0000 UTC m=+1464.492827810" lastFinishedPulling="2025-12-02 10:43:32.399458841 +0000 UTC m=+1465.729597741" observedRunningTime="2025-12-02 10:43:33.262964728 +0000 UTC m=+1466.593103598" watchObservedRunningTime="2025-12-02 10:43:33.274925304 +0000 UTC m=+1466.605064174" Dec 02 10:43:36 crc kubenswrapper[4679]: I1202 10:43:36.278361 4679 generic.go:334] "Generic (PLEG): container finished" podID="e5d79c8f-5612-4597-ba1a-51108bf84927" containerID="4ff9a2c73d0ac0a18f2a3b4dbdcb02c25285adf1b7d715e0efc82afaea777821" exitCode=0 Dec 02 10:43:36 crc kubenswrapper[4679]: I1202 10:43:36.278704 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" event={"ID":"e5d79c8f-5612-4597-ba1a-51108bf84927","Type":"ContainerDied","Data":"4ff9a2c73d0ac0a18f2a3b4dbdcb02c25285adf1b7d715e0efc82afaea777821"} Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.685830 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.754757 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8klr\" (UniqueName: \"kubernetes.io/projected/e5d79c8f-5612-4597-ba1a-51108bf84927-kube-api-access-s8klr\") pod \"e5d79c8f-5612-4597-ba1a-51108bf84927\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.754923 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-ssh-key\") pod \"e5d79c8f-5612-4597-ba1a-51108bf84927\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.755047 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-inventory\") pod \"e5d79c8f-5612-4597-ba1a-51108bf84927\" (UID: \"e5d79c8f-5612-4597-ba1a-51108bf84927\") " Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.765678 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5d79c8f-5612-4597-ba1a-51108bf84927-kube-api-access-s8klr" (OuterVolumeSpecName: "kube-api-access-s8klr") pod "e5d79c8f-5612-4597-ba1a-51108bf84927" (UID: "e5d79c8f-5612-4597-ba1a-51108bf84927"). InnerVolumeSpecName "kube-api-access-s8klr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.781766 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5d79c8f-5612-4597-ba1a-51108bf84927" (UID: "e5d79c8f-5612-4597-ba1a-51108bf84927"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.782654 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-inventory" (OuterVolumeSpecName: "inventory") pod "e5d79c8f-5612-4597-ba1a-51108bf84927" (UID: "e5d79c8f-5612-4597-ba1a-51108bf84927"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.857667 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.857702 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5d79c8f-5612-4597-ba1a-51108bf84927-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:37 crc kubenswrapper[4679]: I1202 10:43:37.857717 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8klr\" (UniqueName: \"kubernetes.io/projected/e5d79c8f-5612-4597-ba1a-51108bf84927-kube-api-access-s8klr\") on node \"crc\" DevicePath \"\"" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.295478 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" event={"ID":"e5d79c8f-5612-4597-ba1a-51108bf84927","Type":"ContainerDied","Data":"023f46fe49531e9828f0d4dbc9cc7d5c4ccbcaaea379f30ea72fb9f6b47bdae0"} Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.295525 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="023f46fe49531e9828f0d4dbc9cc7d5c4ccbcaaea379f30ea72fb9f6b47bdae0" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.295535 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-bzv4b" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.372570 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw"] Dec 02 10:43:38 crc kubenswrapper[4679]: E1202 10:43:38.373150 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5d79c8f-5612-4597-ba1a-51108bf84927" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.373176 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5d79c8f-5612-4597-ba1a-51108bf84927" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.373429 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5d79c8f-5612-4597-ba1a-51108bf84927" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.374247 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.377084 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.377163 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.377755 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.378866 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.384709 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw"] Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.468050 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.468116 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljpx8\" (UniqueName: \"kubernetes.io/projected/dd93c8e7-6ac8-4061-a426-43433fef0376-kube-api-access-ljpx8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.468164 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.468322 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.570114 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.570237 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.570968 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.571023 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljpx8\" (UniqueName: \"kubernetes.io/projected/dd93c8e7-6ac8-4061-a426-43433fef0376-kube-api-access-ljpx8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.574833 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.582425 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.584165 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.596991 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljpx8\" (UniqueName: \"kubernetes.io/projected/dd93c8e7-6ac8-4061-a426-43433fef0376-kube-api-access-ljpx8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:38 crc kubenswrapper[4679]: I1202 10:43:38.691493 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:43:39 crc kubenswrapper[4679]: I1202 10:43:39.281087 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw"] Dec 02 10:43:39 crc kubenswrapper[4679]: I1202 10:43:39.307175 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" event={"ID":"dd93c8e7-6ac8-4061-a426-43433fef0376","Type":"ContainerStarted","Data":"4302ccb63b49ef2a39619a63880d9ee8b05a3e2abbd2616db3e8e41d83247c14"} Dec 02 10:43:40 crc kubenswrapper[4679]: I1202 10:43:40.332257 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" event={"ID":"dd93c8e7-6ac8-4061-a426-43433fef0376","Type":"ContainerStarted","Data":"ce84d6771cfc68376f2ce51ca053c4a0261be395bd1d3cc62e6e263ddfee0142"} Dec 02 10:43:40 crc kubenswrapper[4679]: I1202 10:43:40.365608 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" podStartSLOduration=1.8609224599999998 podStartE2EDuration="2.365589809s" podCreationTimestamp="2025-12-02 10:43:38 +0000 UTC" firstStartedPulling="2025-12-02 10:43:39.280564197 +0000 UTC m=+1472.610703057" lastFinishedPulling="2025-12-02 10:43:39.785231556 +0000 UTC m=+1473.115370406" observedRunningTime="2025-12-02 10:43:40.353258113 +0000 UTC m=+1473.683396973" watchObservedRunningTime="2025-12-02 10:43:40.365589809 +0000 UTC m=+1473.695728669" Dec 02 10:43:46 crc kubenswrapper[4679]: I1202 10:43:46.931799 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:43:46 crc kubenswrapper[4679]: I1202 10:43:46.932397 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:43:46 crc kubenswrapper[4679]: I1202 10:43:46.932439 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:43:46 crc kubenswrapper[4679]: I1202 10:43:46.932962 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0891ad67c232e0d01b60d8bf4be071ef0e56a61b79e4a4f52127a08c972b2d9e"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:43:46 crc kubenswrapper[4679]: I1202 10:43:46.933015 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://0891ad67c232e0d01b60d8bf4be071ef0e56a61b79e4a4f52127a08c972b2d9e" gracePeriod=600 Dec 02 10:43:47 crc kubenswrapper[4679]: I1202 10:43:47.405619 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="0891ad67c232e0d01b60d8bf4be071ef0e56a61b79e4a4f52127a08c972b2d9e" exitCode=0 Dec 02 10:43:47 crc kubenswrapper[4679]: I1202 10:43:47.405682 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"0891ad67c232e0d01b60d8bf4be071ef0e56a61b79e4a4f52127a08c972b2d9e"} Dec 02 10:43:47 crc kubenswrapper[4679]: I1202 10:43:47.405955 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a"} Dec 02 10:43:47 crc kubenswrapper[4679]: I1202 10:43:47.405974 4679 scope.go:117] "RemoveContainer" containerID="1c09445f76d33fdda8f22af126b3018e23a8b609e3131d69d4172f788feafb09" Dec 02 10:44:20 crc kubenswrapper[4679]: I1202 10:44:20.577156 4679 scope.go:117] "RemoveContainer" containerID="ddd15b0d3ba3c2197cf08f8a27e619215b6e7c5bd344af1a6861a46d7adca56c" Dec 02 10:44:20 crc kubenswrapper[4679]: I1202 10:44:20.612231 4679 scope.go:117] "RemoveContainer" containerID="81917a8d774a490d5fa6c3c95d1d994ec418a92049e6aeaafe5bf801b6794d0f" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.145060 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4"] Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.148259 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.151360 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.151383 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.160915 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4"] Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.317831 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfh5w\" (UniqueName: \"kubernetes.io/projected/2ffd70a1-113d-434e-a0e7-701174046fff-kube-api-access-pfh5w\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.317890 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ffd70a1-113d-434e-a0e7-701174046fff-config-volume\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.317937 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ffd70a1-113d-434e-a0e7-701174046fff-secret-volume\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.419751 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfh5w\" (UniqueName: \"kubernetes.io/projected/2ffd70a1-113d-434e-a0e7-701174046fff-kube-api-access-pfh5w\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.419807 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ffd70a1-113d-434e-a0e7-701174046fff-config-volume\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.419843 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ffd70a1-113d-434e-a0e7-701174046fff-secret-volume\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.420830 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ffd70a1-113d-434e-a0e7-701174046fff-config-volume\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.426117 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ffd70a1-113d-434e-a0e7-701174046fff-secret-volume\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.438129 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfh5w\" (UniqueName: \"kubernetes.io/projected/2ffd70a1-113d-434e-a0e7-701174046fff-kube-api-access-pfh5w\") pod \"collect-profiles-29411205-vk9m4\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.479798 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:00 crc kubenswrapper[4679]: I1202 10:45:00.953020 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4"] Dec 02 10:45:01 crc kubenswrapper[4679]: I1202 10:45:01.183756 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" event={"ID":"2ffd70a1-113d-434e-a0e7-701174046fff","Type":"ContainerStarted","Data":"548ad7a48d718656e5f3a32292a4c0e347e15b47c1528b1e52297780c9a00f47"} Dec 02 10:45:02 crc kubenswrapper[4679]: I1202 10:45:02.193722 4679 generic.go:334] "Generic (PLEG): container finished" podID="2ffd70a1-113d-434e-a0e7-701174046fff" containerID="5539149bad511176128adca744e65383edf5fbea0ef6e075d47254a834e64153" exitCode=0 Dec 02 10:45:02 crc kubenswrapper[4679]: I1202 10:45:02.194585 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" event={"ID":"2ffd70a1-113d-434e-a0e7-701174046fff","Type":"ContainerDied","Data":"5539149bad511176128adca744e65383edf5fbea0ef6e075d47254a834e64153"} Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.541417 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.677497 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfh5w\" (UniqueName: \"kubernetes.io/projected/2ffd70a1-113d-434e-a0e7-701174046fff-kube-api-access-pfh5w\") pod \"2ffd70a1-113d-434e-a0e7-701174046fff\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.677757 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ffd70a1-113d-434e-a0e7-701174046fff-config-volume\") pod \"2ffd70a1-113d-434e-a0e7-701174046fff\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.677794 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ffd70a1-113d-434e-a0e7-701174046fff-secret-volume\") pod \"2ffd70a1-113d-434e-a0e7-701174046fff\" (UID: \"2ffd70a1-113d-434e-a0e7-701174046fff\") " Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.678585 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ffd70a1-113d-434e-a0e7-701174046fff-config-volume" (OuterVolumeSpecName: "config-volume") pod "2ffd70a1-113d-434e-a0e7-701174046fff" (UID: "2ffd70a1-113d-434e-a0e7-701174046fff"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.683774 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ffd70a1-113d-434e-a0e7-701174046fff-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2ffd70a1-113d-434e-a0e7-701174046fff" (UID: "2ffd70a1-113d-434e-a0e7-701174046fff"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.683874 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ffd70a1-113d-434e-a0e7-701174046fff-kube-api-access-pfh5w" (OuterVolumeSpecName: "kube-api-access-pfh5w") pod "2ffd70a1-113d-434e-a0e7-701174046fff" (UID: "2ffd70a1-113d-434e-a0e7-701174046fff"). InnerVolumeSpecName "kube-api-access-pfh5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.779612 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ffd70a1-113d-434e-a0e7-701174046fff-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.779653 4679 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ffd70a1-113d-434e-a0e7-701174046fff-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 10:45:03 crc kubenswrapper[4679]: I1202 10:45:03.779667 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfh5w\" (UniqueName: \"kubernetes.io/projected/2ffd70a1-113d-434e-a0e7-701174046fff-kube-api-access-pfh5w\") on node \"crc\" DevicePath \"\"" Dec 02 10:45:04 crc kubenswrapper[4679]: I1202 10:45:04.215933 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" event={"ID":"2ffd70a1-113d-434e-a0e7-701174046fff","Type":"ContainerDied","Data":"548ad7a48d718656e5f3a32292a4c0e347e15b47c1528b1e52297780c9a00f47"} Dec 02 10:45:04 crc kubenswrapper[4679]: I1202 10:45:04.216236 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="548ad7a48d718656e5f3a32292a4c0e347e15b47c1528b1e52297780c9a00f47" Dec 02 10:45:04 crc kubenswrapper[4679]: I1202 10:45:04.216005 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4" Dec 02 10:45:20 crc kubenswrapper[4679]: I1202 10:45:20.727669 4679 scope.go:117] "RemoveContainer" containerID="16c9ce7605d2726c95ebb815b0034375cb0e3f2330bfb6f4df9096a1f428fa2f" Dec 02 10:45:20 crc kubenswrapper[4679]: I1202 10:45:20.751442 4679 scope.go:117] "RemoveContainer" containerID="40bd0ffa15a8f1c691c47a4c9d9e744e3a368de748fe3021427a268baebdb592" Dec 02 10:46:16 crc kubenswrapper[4679]: I1202 10:46:16.931881 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:46:16 crc kubenswrapper[4679]: I1202 10:46:16.933673 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.315370 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b2xjv"] Dec 02 10:46:21 crc kubenswrapper[4679]: E1202 10:46:21.316627 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ffd70a1-113d-434e-a0e7-701174046fff" containerName="collect-profiles" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.316648 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ffd70a1-113d-434e-a0e7-701174046fff" containerName="collect-profiles" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.316890 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ffd70a1-113d-434e-a0e7-701174046fff" containerName="collect-profiles" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.318493 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.343453 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b2xjv"] Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.420722 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp2nw\" (UniqueName: \"kubernetes.io/projected/41a03b9a-9119-474c-9ae4-725879cbc78e-kube-api-access-vp2nw\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.420767 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41a03b9a-9119-474c-9ae4-725879cbc78e-utilities\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.420836 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41a03b9a-9119-474c-9ae4-725879cbc78e-catalog-content\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.522728 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp2nw\" (UniqueName: \"kubernetes.io/projected/41a03b9a-9119-474c-9ae4-725879cbc78e-kube-api-access-vp2nw\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.522812 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41a03b9a-9119-474c-9ae4-725879cbc78e-utilities\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.522917 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41a03b9a-9119-474c-9ae4-725879cbc78e-catalog-content\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.523383 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41a03b9a-9119-474c-9ae4-725879cbc78e-catalog-content\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.523640 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41a03b9a-9119-474c-9ae4-725879cbc78e-utilities\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.551943 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp2nw\" (UniqueName: \"kubernetes.io/projected/41a03b9a-9119-474c-9ae4-725879cbc78e-kube-api-access-vp2nw\") pod \"community-operators-b2xjv\" (UID: \"41a03b9a-9119-474c-9ae4-725879cbc78e\") " pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:21 crc kubenswrapper[4679]: I1202 10:46:21.636552 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:22 crc kubenswrapper[4679]: I1202 10:46:22.244534 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b2xjv"] Dec 02 10:46:23 crc kubenswrapper[4679]: I1202 10:46:23.148685 4679 generic.go:334] "Generic (PLEG): container finished" podID="41a03b9a-9119-474c-9ae4-725879cbc78e" containerID="08882890baa258400e85aa4c3bd20cf85085ffdd67360800c2cb1390ed24feda" exitCode=0 Dec 02 10:46:23 crc kubenswrapper[4679]: I1202 10:46:23.148789 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2xjv" event={"ID":"41a03b9a-9119-474c-9ae4-725879cbc78e","Type":"ContainerDied","Data":"08882890baa258400e85aa4c3bd20cf85085ffdd67360800c2cb1390ed24feda"} Dec 02 10:46:23 crc kubenswrapper[4679]: I1202 10:46:23.148972 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2xjv" event={"ID":"41a03b9a-9119-474c-9ae4-725879cbc78e","Type":"ContainerStarted","Data":"eb097ae5161139476b6f8cc7bbcd8dee7093699a9a9b1b9b3cb0cb9c0fe658aa"} Dec 02 10:46:28 crc kubenswrapper[4679]: I1202 10:46:28.201706 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2xjv" event={"ID":"41a03b9a-9119-474c-9ae4-725879cbc78e","Type":"ContainerStarted","Data":"4e22815b7946a75a191092d591957573f1f6f3ba2e5ca226ed6eb9e97c896b6c"} Dec 02 10:46:29 crc kubenswrapper[4679]: I1202 10:46:29.214113 4679 generic.go:334] "Generic (PLEG): container finished" podID="41a03b9a-9119-474c-9ae4-725879cbc78e" containerID="4e22815b7946a75a191092d591957573f1f6f3ba2e5ca226ed6eb9e97c896b6c" exitCode=0 Dec 02 10:46:29 crc kubenswrapper[4679]: I1202 10:46:29.214154 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2xjv" event={"ID":"41a03b9a-9119-474c-9ae4-725879cbc78e","Type":"ContainerDied","Data":"4e22815b7946a75a191092d591957573f1f6f3ba2e5ca226ed6eb9e97c896b6c"} Dec 02 10:46:29 crc kubenswrapper[4679]: I1202 10:46:29.217132 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:46:30 crc kubenswrapper[4679]: I1202 10:46:30.247619 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2xjv" event={"ID":"41a03b9a-9119-474c-9ae4-725879cbc78e","Type":"ContainerStarted","Data":"efc0c02fdf2e18f1b831783493f4190983744339f13f896c92bbd2f724369173"} Dec 02 10:46:30 crc kubenswrapper[4679]: I1202 10:46:30.273890 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b2xjv" podStartSLOduration=2.6779795650000002 podStartE2EDuration="9.273874727s" podCreationTimestamp="2025-12-02 10:46:21 +0000 UTC" firstStartedPulling="2025-12-02 10:46:23.151122566 +0000 UTC m=+1636.481261426" lastFinishedPulling="2025-12-02 10:46:29.747017728 +0000 UTC m=+1643.077156588" observedRunningTime="2025-12-02 10:46:30.268767749 +0000 UTC m=+1643.598906609" watchObservedRunningTime="2025-12-02 10:46:30.273874727 +0000 UTC m=+1643.604013577" Dec 02 10:46:31 crc kubenswrapper[4679]: I1202 10:46:31.637711 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:31 crc kubenswrapper[4679]: I1202 10:46:31.638001 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:32 crc kubenswrapper[4679]: I1202 10:46:32.700406 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-b2xjv" podUID="41a03b9a-9119-474c-9ae4-725879cbc78e" containerName="registry-server" probeResult="failure" output=< Dec 02 10:46:32 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 10:46:32 crc kubenswrapper[4679]: > Dec 02 10:46:41 crc kubenswrapper[4679]: I1202 10:46:41.691937 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:41 crc kubenswrapper[4679]: I1202 10:46:41.757266 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b2xjv" Dec 02 10:46:41 crc kubenswrapper[4679]: I1202 10:46:41.834392 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b2xjv"] Dec 02 10:46:41 crc kubenswrapper[4679]: I1202 10:46:41.943118 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76xkq"] Dec 02 10:46:41 crc kubenswrapper[4679]: I1202 10:46:41.943584 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-76xkq" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="registry-server" containerID="cri-o://2eece58ddecd8fdf638aa95f1cef9562ab4dc6d6253b12652d517cd54500cfb4" gracePeriod=2 Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.355514 4679 generic.go:334] "Generic (PLEG): container finished" podID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerID="2eece58ddecd8fdf638aa95f1cef9562ab4dc6d6253b12652d517cd54500cfb4" exitCode=0 Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.355719 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerDied","Data":"2eece58ddecd8fdf638aa95f1cef9562ab4dc6d6253b12652d517cd54500cfb4"} Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.355850 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76xkq" event={"ID":"5610cf13-1e22-4bb5-a5da-3b67207aef40","Type":"ContainerDied","Data":"c7d8dc0c700c9d33ef1898f99133c27d41257118164af4b9b5aa22d6e1e9c2f2"} Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.355866 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7d8dc0c700c9d33ef1898f99133c27d41257118164af4b9b5aa22d6e1e9c2f2" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.437524 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.546222 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-utilities\") pod \"5610cf13-1e22-4bb5-a5da-3b67207aef40\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.546530 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcb82\" (UniqueName: \"kubernetes.io/projected/5610cf13-1e22-4bb5-a5da-3b67207aef40-kube-api-access-xcb82\") pod \"5610cf13-1e22-4bb5-a5da-3b67207aef40\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.546655 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-catalog-content\") pod \"5610cf13-1e22-4bb5-a5da-3b67207aef40\" (UID: \"5610cf13-1e22-4bb5-a5da-3b67207aef40\") " Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.546868 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-utilities" (OuterVolumeSpecName: "utilities") pod "5610cf13-1e22-4bb5-a5da-3b67207aef40" (UID: "5610cf13-1e22-4bb5-a5da-3b67207aef40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.547677 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.554664 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5610cf13-1e22-4bb5-a5da-3b67207aef40-kube-api-access-xcb82" (OuterVolumeSpecName: "kube-api-access-xcb82") pod "5610cf13-1e22-4bb5-a5da-3b67207aef40" (UID: "5610cf13-1e22-4bb5-a5da-3b67207aef40"). InnerVolumeSpecName "kube-api-access-xcb82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.593025 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5610cf13-1e22-4bb5-a5da-3b67207aef40" (UID: "5610cf13-1e22-4bb5-a5da-3b67207aef40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.650160 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcb82\" (UniqueName: \"kubernetes.io/projected/5610cf13-1e22-4bb5-a5da-3b67207aef40-kube-api-access-xcb82\") on node \"crc\" DevicePath \"\"" Dec 02 10:46:42 crc kubenswrapper[4679]: I1202 10:46:42.650195 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5610cf13-1e22-4bb5-a5da-3b67207aef40-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:46:43 crc kubenswrapper[4679]: I1202 10:46:43.364620 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76xkq" Dec 02 10:46:43 crc kubenswrapper[4679]: I1202 10:46:43.392510 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76xkq"] Dec 02 10:46:43 crc kubenswrapper[4679]: I1202 10:46:43.400089 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-76xkq"] Dec 02 10:46:44 crc kubenswrapper[4679]: I1202 10:46:44.923690 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" path="/var/lib/kubelet/pods/5610cf13-1e22-4bb5-a5da-3b67207aef40/volumes" Dec 02 10:46:46 crc kubenswrapper[4679]: I1202 10:46:46.931941 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:46:46 crc kubenswrapper[4679]: I1202 10:46:46.933068 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:46:58 crc kubenswrapper[4679]: I1202 10:46:58.501862 4679 generic.go:334] "Generic (PLEG): container finished" podID="dd93c8e7-6ac8-4061-a426-43433fef0376" containerID="ce84d6771cfc68376f2ce51ca053c4a0261be395bd1d3cc62e6e263ddfee0142" exitCode=0 Dec 02 10:46:58 crc kubenswrapper[4679]: I1202 10:46:58.502492 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" event={"ID":"dd93c8e7-6ac8-4061-a426-43433fef0376","Type":"ContainerDied","Data":"ce84d6771cfc68376f2ce51ca053c4a0261be395bd1d3cc62e6e263ddfee0142"} Dec 02 10:46:59 crc kubenswrapper[4679]: I1202 10:46:59.953257 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.105681 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljpx8\" (UniqueName: \"kubernetes.io/projected/dd93c8e7-6ac8-4061-a426-43433fef0376-kube-api-access-ljpx8\") pod \"dd93c8e7-6ac8-4061-a426-43433fef0376\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.105864 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-bootstrap-combined-ca-bundle\") pod \"dd93c8e7-6ac8-4061-a426-43433fef0376\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.105967 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-ssh-key\") pod \"dd93c8e7-6ac8-4061-a426-43433fef0376\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.106064 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-inventory\") pod \"dd93c8e7-6ac8-4061-a426-43433fef0376\" (UID: \"dd93c8e7-6ac8-4061-a426-43433fef0376\") " Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.115570 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "dd93c8e7-6ac8-4061-a426-43433fef0376" (UID: "dd93c8e7-6ac8-4061-a426-43433fef0376"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.115723 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd93c8e7-6ac8-4061-a426-43433fef0376-kube-api-access-ljpx8" (OuterVolumeSpecName: "kube-api-access-ljpx8") pod "dd93c8e7-6ac8-4061-a426-43433fef0376" (UID: "dd93c8e7-6ac8-4061-a426-43433fef0376"). InnerVolumeSpecName "kube-api-access-ljpx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.144093 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-inventory" (OuterVolumeSpecName: "inventory") pod "dd93c8e7-6ac8-4061-a426-43433fef0376" (UID: "dd93c8e7-6ac8-4061-a426-43433fef0376"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.146557 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd93c8e7-6ac8-4061-a426-43433fef0376" (UID: "dd93c8e7-6ac8-4061-a426-43433fef0376"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.209663 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.210111 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljpx8\" (UniqueName: \"kubernetes.io/projected/dd93c8e7-6ac8-4061-a426-43433fef0376-kube-api-access-ljpx8\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.210283 4679 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.210429 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd93c8e7-6ac8-4061-a426-43433fef0376-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.529854 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" event={"ID":"dd93c8e7-6ac8-4061-a426-43433fef0376","Type":"ContainerDied","Data":"4302ccb63b49ef2a39619a63880d9ee8b05a3e2abbd2616db3e8e41d83247c14"} Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.530204 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4302ccb63b49ef2a39619a63880d9ee8b05a3e2abbd2616db3e8e41d83247c14" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.529987 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.630191 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2"] Dec 02 10:47:00 crc kubenswrapper[4679]: E1202 10:47:00.630705 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="extract-content" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.630730 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="extract-content" Dec 02 10:47:00 crc kubenswrapper[4679]: E1202 10:47:00.630764 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="registry-server" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.630772 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="registry-server" Dec 02 10:47:00 crc kubenswrapper[4679]: E1202 10:47:00.630785 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="extract-utilities" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.630794 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="extract-utilities" Dec 02 10:47:00 crc kubenswrapper[4679]: E1202 10:47:00.630809 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd93c8e7-6ac8-4061-a426-43433fef0376" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.630817 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd93c8e7-6ac8-4061-a426-43433fef0376" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.631042 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5610cf13-1e22-4bb5-a5da-3b67207aef40" containerName="registry-server" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.631070 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd93c8e7-6ac8-4061-a426-43433fef0376" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.631898 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.635067 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.635233 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.635137 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.639401 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.643464 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2"] Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.718644 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdhl2\" (UniqueName: \"kubernetes.io/projected/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-kube-api-access-bdhl2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.718740 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.718799 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.821082 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdhl2\" (UniqueName: \"kubernetes.io/projected/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-kube-api-access-bdhl2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.821184 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.821250 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.824712 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.832941 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.844069 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdhl2\" (UniqueName: \"kubernetes.io/projected/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-kube-api-access-bdhl2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:00 crc kubenswrapper[4679]: I1202 10:47:00.952577 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:47:01 crc kubenswrapper[4679]: I1202 10:47:01.515593 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2"] Dec 02 10:47:01 crc kubenswrapper[4679]: W1202 10:47:01.520832 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2119b0d5_f21e_4694_9fc8_7bf3bed24ecc.slice/crio-838af880d0731124279ca93c537fa2cdb95d7eced021338a414f8b49ae0b6e0a WatchSource:0}: Error finding container 838af880d0731124279ca93c537fa2cdb95d7eced021338a414f8b49ae0b6e0a: Status 404 returned error can't find the container with id 838af880d0731124279ca93c537fa2cdb95d7eced021338a414f8b49ae0b6e0a Dec 02 10:47:01 crc kubenswrapper[4679]: I1202 10:47:01.543569 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" event={"ID":"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc","Type":"ContainerStarted","Data":"838af880d0731124279ca93c537fa2cdb95d7eced021338a414f8b49ae0b6e0a"} Dec 02 10:47:02 crc kubenswrapper[4679]: I1202 10:47:02.553501 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" event={"ID":"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc","Type":"ContainerStarted","Data":"d452b43f8951f013b8fbcc6a31a1ed0b49f25fd88e9f84cdeb4ab19d4ffdf808"} Dec 02 10:47:02 crc kubenswrapper[4679]: I1202 10:47:02.587018 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" podStartSLOduration=2.090573546 podStartE2EDuration="2.586995906s" podCreationTimestamp="2025-12-02 10:47:00 +0000 UTC" firstStartedPulling="2025-12-02 10:47:01.525970138 +0000 UTC m=+1674.856109008" lastFinishedPulling="2025-12-02 10:47:02.022392508 +0000 UTC m=+1675.352531368" observedRunningTime="2025-12-02 10:47:02.579894521 +0000 UTC m=+1675.910033381" watchObservedRunningTime="2025-12-02 10:47:02.586995906 +0000 UTC m=+1675.917134766" Dec 02 10:47:16 crc kubenswrapper[4679]: I1202 10:47:16.932080 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:47:16 crc kubenswrapper[4679]: I1202 10:47:16.932812 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:47:16 crc kubenswrapper[4679]: I1202 10:47:16.932870 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:47:16 crc kubenswrapper[4679]: I1202 10:47:16.933825 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:47:16 crc kubenswrapper[4679]: I1202 10:47:16.933915 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" gracePeriod=600 Dec 02 10:47:17 crc kubenswrapper[4679]: E1202 10:47:17.065636 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:47:17 crc kubenswrapper[4679]: I1202 10:47:17.701934 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" exitCode=0 Dec 02 10:47:17 crc kubenswrapper[4679]: I1202 10:47:17.701995 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a"} Dec 02 10:47:17 crc kubenswrapper[4679]: I1202 10:47:17.702263 4679 scope.go:117] "RemoveContainer" containerID="0891ad67c232e0d01b60d8bf4be071ef0e56a61b79e4a4f52127a08c972b2d9e" Dec 02 10:47:17 crc kubenswrapper[4679]: I1202 10:47:17.702933 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:47:17 crc kubenswrapper[4679]: E1202 10:47:17.703346 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:47:20 crc kubenswrapper[4679]: I1202 10:47:20.850022 4679 scope.go:117] "RemoveContainer" containerID="2eece58ddecd8fdf638aa95f1cef9562ab4dc6d6253b12652d517cd54500cfb4" Dec 02 10:47:20 crc kubenswrapper[4679]: I1202 10:47:20.885516 4679 scope.go:117] "RemoveContainer" containerID="46370d8c7e8e7dc32df22aed0891d3fc00c62705886a09f6f53b5c4583ebbcdf" Dec 02 10:47:20 crc kubenswrapper[4679]: I1202 10:47:20.914892 4679 scope.go:117] "RemoveContainer" containerID="09a5370e9d2bb3ccead6f3ab1604562f5c2d5842e808a189eba8836efb7f28d0" Dec 02 10:47:20 crc kubenswrapper[4679]: I1202 10:47:20.968653 4679 scope.go:117] "RemoveContainer" containerID="51fbfe9ca51bb63f94d28d8e629804c8b943dbeee498ec2291f9894d9975073c" Dec 02 10:47:20 crc kubenswrapper[4679]: I1202 10:47:20.985214 4679 scope.go:117] "RemoveContainer" containerID="f5f7c2604ac78f4621b37ef7fec64c1841a2d5aa617fd0042a959fddb22c6b2f" Dec 02 10:47:27 crc kubenswrapper[4679]: I1202 10:47:27.909251 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:47:27 crc kubenswrapper[4679]: E1202 10:47:27.909917 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.590945 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lknlr"] Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.594278 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.607812 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lknlr"] Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.682428 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-utilities\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.682516 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-289kc\" (UniqueName: \"kubernetes.io/projected/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-kube-api-access-289kc\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.682612 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-catalog-content\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.784036 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-catalog-content\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.784115 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-utilities\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.784188 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-289kc\" (UniqueName: \"kubernetes.io/projected/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-kube-api-access-289kc\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.784624 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-catalog-content\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.784714 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-utilities\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.805957 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-289kc\" (UniqueName: \"kubernetes.io/projected/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-kube-api-access-289kc\") pod \"certified-operators-lknlr\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:28 crc kubenswrapper[4679]: I1202 10:47:28.915661 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:29 crc kubenswrapper[4679]: I1202 10:47:29.451266 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lknlr"] Dec 02 10:47:29 crc kubenswrapper[4679]: I1202 10:47:29.844596 4679 generic.go:334] "Generic (PLEG): container finished" podID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerID="5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65" exitCode=0 Dec 02 10:47:29 crc kubenswrapper[4679]: I1202 10:47:29.844652 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerDied","Data":"5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65"} Dec 02 10:47:29 crc kubenswrapper[4679]: I1202 10:47:29.844699 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerStarted","Data":"2598a8be283d625d578abbfd5b79d669eac2dfba83fc766ef9c8a2db41851ac7"} Dec 02 10:47:30 crc kubenswrapper[4679]: I1202 10:47:30.853186 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerStarted","Data":"25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df"} Dec 02 10:47:31 crc kubenswrapper[4679]: I1202 10:47:31.861263 4679 generic.go:334] "Generic (PLEG): container finished" podID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerID="25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df" exitCode=0 Dec 02 10:47:31 crc kubenswrapper[4679]: I1202 10:47:31.861328 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerDied","Data":"25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df"} Dec 02 10:47:33 crc kubenswrapper[4679]: I1202 10:47:33.905291 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lknlr" podStartSLOduration=2.07518475 podStartE2EDuration="5.905272811s" podCreationTimestamp="2025-12-02 10:47:28 +0000 UTC" firstStartedPulling="2025-12-02 10:47:29.847654279 +0000 UTC m=+1703.177793139" lastFinishedPulling="2025-12-02 10:47:33.67774234 +0000 UTC m=+1707.007881200" observedRunningTime="2025-12-02 10:47:33.900116203 +0000 UTC m=+1707.230255063" watchObservedRunningTime="2025-12-02 10:47:33.905272811 +0000 UTC m=+1707.235411671" Dec 02 10:47:34 crc kubenswrapper[4679]: I1202 10:47:34.894434 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerStarted","Data":"2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0"} Dec 02 10:47:38 crc kubenswrapper[4679]: I1202 10:47:38.932046 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:38 crc kubenswrapper[4679]: I1202 10:47:38.932591 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:38 crc kubenswrapper[4679]: I1202 10:47:38.977359 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:39 crc kubenswrapper[4679]: I1202 10:47:39.984630 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:40 crc kubenswrapper[4679]: I1202 10:47:40.037482 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lknlr"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.046853 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-164a-account-create-update-bsvts"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.055875 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-w6dwk"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.063975 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-w84l8"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.072912 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-2xs5t"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.087190 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e4cf-account-create-update-fp8rh"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.098208 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-w6dwk"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.110166 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-164a-account-create-update-bsvts"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.121708 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-2xs5t"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.133225 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-w84l8"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.141726 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e4cf-account-create-update-fp8rh"] Dec 02 10:47:41 crc kubenswrapper[4679]: I1202 10:47:41.969339 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lknlr" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="registry-server" containerID="cri-o://2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0" gracePeriod=2 Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.035592 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-bf39-account-create-update-9gvms"] Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.044729 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-bf39-account-create-update-9gvms"] Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.408859 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.581460 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-catalog-content\") pod \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.581536 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-289kc\" (UniqueName: \"kubernetes.io/projected/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-kube-api-access-289kc\") pod \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.581679 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-utilities\") pod \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\" (UID: \"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5\") " Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.582484 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-utilities" (OuterVolumeSpecName: "utilities") pod "082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" (UID: "082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.590008 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-kube-api-access-289kc" (OuterVolumeSpecName: "kube-api-access-289kc") pod "082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" (UID: "082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5"). InnerVolumeSpecName "kube-api-access-289kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.635024 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" (UID: "082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.683662 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.683691 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.683707 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-289kc\" (UniqueName: \"kubernetes.io/projected/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5-kube-api-access-289kc\") on node \"crc\" DevicePath \"\"" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.909259 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:47:42 crc kubenswrapper[4679]: E1202 10:47:42.910050 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.931253 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a6ecd7-a808-43a6-9562-953818cc9020" path="/var/lib/kubelet/pods/00a6ecd7-a808-43a6-9562-953818cc9020/volumes" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.932806 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59eae751-1d2f-4ce3-a386-d2dcc0727717" path="/var/lib/kubelet/pods/59eae751-1d2f-4ce3-a386-d2dcc0727717/volumes" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.934333 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2f792c0-0b54-4599-aea3-d5567072e6f7" path="/var/lib/kubelet/pods/c2f792c0-0b54-4599-aea3-d5567072e6f7/volumes" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.935756 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db8eb5a7-5fef-4f4a-905f-40103b0f513d" path="/var/lib/kubelet/pods/db8eb5a7-5fef-4f4a-905f-40103b0f513d/volumes" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.938041 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6f8df03-1845-4c2b-a30a-3d8186287f73" path="/var/lib/kubelet/pods/e6f8df03-1845-4c2b-a30a-3d8186287f73/volumes" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.938701 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5c4d5af-bdac-4486-b0ab-f389cd3331bb" path="/var/lib/kubelet/pods/f5c4d5af-bdac-4486-b0ab-f389cd3331bb/volumes" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.982700 4679 generic.go:334] "Generic (PLEG): container finished" podID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerID="2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0" exitCode=0 Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.982761 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerDied","Data":"2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0"} Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.982828 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lknlr" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.982867 4679 scope.go:117] "RemoveContainer" containerID="2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0" Dec 02 10:47:42 crc kubenswrapper[4679]: I1202 10:47:42.982845 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lknlr" event={"ID":"082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5","Type":"ContainerDied","Data":"2598a8be283d625d578abbfd5b79d669eac2dfba83fc766ef9c8a2db41851ac7"} Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.013431 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lknlr"] Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.021378 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lknlr"] Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.025405 4679 scope.go:117] "RemoveContainer" containerID="25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.050870 4679 scope.go:117] "RemoveContainer" containerID="5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.118528 4679 scope.go:117] "RemoveContainer" containerID="2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0" Dec 02 10:47:43 crc kubenswrapper[4679]: E1202 10:47:43.118952 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0\": container with ID starting with 2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0 not found: ID does not exist" containerID="2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.119019 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0"} err="failed to get container status \"2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0\": rpc error: code = NotFound desc = could not find container \"2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0\": container with ID starting with 2c63290d1c0204c4f9ecd97c59ae52813786c8d2c46fc874820b9f67702c47b0 not found: ID does not exist" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.119061 4679 scope.go:117] "RemoveContainer" containerID="25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df" Dec 02 10:47:43 crc kubenswrapper[4679]: E1202 10:47:43.119385 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df\": container with ID starting with 25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df not found: ID does not exist" containerID="25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.119431 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df"} err="failed to get container status \"25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df\": rpc error: code = NotFound desc = could not find container \"25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df\": container with ID starting with 25c64a8920748287300bcfcaed2fd695d6078fb9f90fb53fdfd4f788914686df not found: ID does not exist" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.119466 4679 scope.go:117] "RemoveContainer" containerID="5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65" Dec 02 10:47:43 crc kubenswrapper[4679]: E1202 10:47:43.119799 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65\": container with ID starting with 5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65 not found: ID does not exist" containerID="5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65" Dec 02 10:47:43 crc kubenswrapper[4679]: I1202 10:47:43.119845 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65"} err="failed to get container status \"5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65\": rpc error: code = NotFound desc = could not find container \"5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65\": container with ID starting with 5493b8d65bf1be083742ee57937ecbdc5a5adede96048e7d874a3208c8757a65 not found: ID does not exist" Dec 02 10:47:44 crc kubenswrapper[4679]: I1202 10:47:44.922702 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" path="/var/lib/kubelet/pods/082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5/volumes" Dec 02 10:47:54 crc kubenswrapper[4679]: I1202 10:47:54.908857 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:47:54 crc kubenswrapper[4679]: E1202 10:47:54.909595 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:48:05 crc kubenswrapper[4679]: I1202 10:48:05.908901 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:48:05 crc kubenswrapper[4679]: E1202 10:48:05.909716 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:48:08 crc kubenswrapper[4679]: I1202 10:48:08.033686 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-mjmmw"] Dec 02 10:48:08 crc kubenswrapper[4679]: I1202 10:48:08.041485 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-mjmmw"] Dec 02 10:48:08 crc kubenswrapper[4679]: I1202 10:48:08.919687 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d57d916-49db-43e8-b1cd-7bd5406d4812" path="/var/lib/kubelet/pods/0d57d916-49db-43e8-b1cd-7bd5406d4812/volumes" Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.054536 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qbg4x"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.065783 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b3a9-account-create-update-jg27q"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.092083 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e045-account-create-update-c77tp"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.099925 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qbg4x"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.107439 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-46e9-account-create-update-dbkf8"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.114511 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e045-account-create-update-c77tp"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.122037 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-ctlv8"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.130795 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b3a9-account-create-update-jg27q"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.139501 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-46e9-account-create-update-dbkf8"] Dec 02 10:48:11 crc kubenswrapper[4679]: I1202 10:48:11.149821 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-ctlv8"] Dec 02 10:48:12 crc kubenswrapper[4679]: I1202 10:48:12.920795 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0969458b-57a6-4ec6-92b9-870ee1cfc6da" path="/var/lib/kubelet/pods/0969458b-57a6-4ec6-92b9-870ee1cfc6da/volumes" Dec 02 10:48:12 crc kubenswrapper[4679]: I1202 10:48:12.921825 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="595fb7b3-d70c-4cf2-98c4-287fff3ec997" path="/var/lib/kubelet/pods/595fb7b3-d70c-4cf2-98c4-287fff3ec997/volumes" Dec 02 10:48:12 crc kubenswrapper[4679]: I1202 10:48:12.922348 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5b4081a-2347-45b3-b5db-07de3d5a90c2" path="/var/lib/kubelet/pods/c5b4081a-2347-45b3-b5db-07de3d5a90c2/volumes" Dec 02 10:48:12 crc kubenswrapper[4679]: I1202 10:48:12.922871 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4c80f2-bd41-4e94-b8e9-c8d6616357e6" path="/var/lib/kubelet/pods/db4c80f2-bd41-4e94-b8e9-c8d6616357e6/volumes" Dec 02 10:48:12 crc kubenswrapper[4679]: I1202 10:48:12.923898 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7f3b5a2-58d0-45cc-9547-8b9eea1a9040" path="/var/lib/kubelet/pods/f7f3b5a2-58d0-45cc-9547-8b9eea1a9040/volumes" Dec 02 10:48:15 crc kubenswrapper[4679]: I1202 10:48:15.039236 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-dk9q8"] Dec 02 10:48:15 crc kubenswrapper[4679]: I1202 10:48:15.054900 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-dk9q8"] Dec 02 10:48:16 crc kubenswrapper[4679]: I1202 10:48:16.929017 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f9bcb09-6b18-4357-b474-0db2b33e1641" path="/var/lib/kubelet/pods/9f9bcb09-6b18-4357-b474-0db2b33e1641/volumes" Dec 02 10:48:18 crc kubenswrapper[4679]: I1202 10:48:18.909055 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:48:18 crc kubenswrapper[4679]: E1202 10:48:18.909612 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.047891 4679 scope.go:117] "RemoveContainer" containerID="e3e3ffd41d2f57357f50ff15e951390ff9b6ef2feefc458118acf3b17dca7b95" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.082106 4679 scope.go:117] "RemoveContainer" containerID="8dae82f6710a97d71a8bb2dd5eb50004498520c0491b19bd114b3263258a730e" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.122687 4679 scope.go:117] "RemoveContainer" containerID="2da94eb5efb39877d43b62507e84d1bea5bc435f7fbd004b02a008dc3d59e28a" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.164004 4679 scope.go:117] "RemoveContainer" containerID="59132a40f0f7bf6fb36a5e0df1bd62d1c7827c4155601e5c562af3e1e31536ec" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.203182 4679 scope.go:117] "RemoveContainer" containerID="b8610e7c12a0b7efd7c39c1944ed1b04aadc67ee634cf5d4dae991e3870e82f4" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.280174 4679 scope.go:117] "RemoveContainer" containerID="c100dfeef4fdb34a70898323eb438a18537e6364c365397f0168d3e0197551f1" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.313220 4679 scope.go:117] "RemoveContainer" containerID="3a6a97813d3d5e88564b30008c92dc07ec16c17d3e6c1a407fe5df9d448f2c9d" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.347790 4679 scope.go:117] "RemoveContainer" containerID="ae341e015d059b83e96e1f4f910fcc9ad91d74d843c59ab7a7035273a6fdd1cc" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.370875 4679 scope.go:117] "RemoveContainer" containerID="424e1d8b153bdce6f42ed6328cdaa50f7c461b3e8bccc32a8b19058b7892596c" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.397009 4679 scope.go:117] "RemoveContainer" containerID="cba53ad1d4dda8482a8887e03b1dd1b51a10ceaf76382f10be8b7df5b866dd4c" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.420552 4679 scope.go:117] "RemoveContainer" containerID="a8d697ea66b8450103078fcdf58a8daa4cb4f27754938a7c0c5db8e41bcec107" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.438723 4679 scope.go:117] "RemoveContainer" containerID="c070b247a78fa29a7919fe681f74d9c5f9764b148e44474d917e3e58c3eb7bd0" Dec 02 10:48:21 crc kubenswrapper[4679]: I1202 10:48:21.456466 4679 scope.go:117] "RemoveContainer" containerID="b7d2a2ad8f04c9cd1576165896296ca93185f6babc3d040bc804ca53afd1d7b3" Dec 02 10:48:33 crc kubenswrapper[4679]: I1202 10:48:33.909270 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:48:33 crc kubenswrapper[4679]: E1202 10:48:33.910270 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:48:40 crc kubenswrapper[4679]: I1202 10:48:40.742741 4679 generic.go:334] "Generic (PLEG): container finished" podID="2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" containerID="d452b43f8951f013b8fbcc6a31a1ed0b49f25fd88e9f84cdeb4ab19d4ffdf808" exitCode=0 Dec 02 10:48:40 crc kubenswrapper[4679]: I1202 10:48:40.742823 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" event={"ID":"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc","Type":"ContainerDied","Data":"d452b43f8951f013b8fbcc6a31a1ed0b49f25fd88e9f84cdeb4ab19d4ffdf808"} Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.130718 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.304078 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdhl2\" (UniqueName: \"kubernetes.io/projected/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-kube-api-access-bdhl2\") pod \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.304239 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-inventory\") pod \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.304462 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-ssh-key\") pod \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\" (UID: \"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc\") " Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.311937 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-kube-api-access-bdhl2" (OuterVolumeSpecName: "kube-api-access-bdhl2") pod "2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" (UID: "2119b0d5-f21e-4694-9fc8-7bf3bed24ecc"). InnerVolumeSpecName "kube-api-access-bdhl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.341530 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" (UID: "2119b0d5-f21e-4694-9fc8-7bf3bed24ecc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.347064 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-inventory" (OuterVolumeSpecName: "inventory") pod "2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" (UID: "2119b0d5-f21e-4694-9fc8-7bf3bed24ecc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.410672 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdhl2\" (UniqueName: \"kubernetes.io/projected/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-kube-api-access-bdhl2\") on node \"crc\" DevicePath \"\"" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.410709 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.410727 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2119b0d5-f21e-4694-9fc8-7bf3bed24ecc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.765436 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" event={"ID":"2119b0d5-f21e-4694-9fc8-7bf3bed24ecc","Type":"ContainerDied","Data":"838af880d0731124279ca93c537fa2cdb95d7eced021338a414f8b49ae0b6e0a"} Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.765845 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="838af880d0731124279ca93c537fa2cdb95d7eced021338a414f8b49ae0b6e0a" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.765548 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.865561 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv"] Dec 02 10:48:42 crc kubenswrapper[4679]: E1202 10:48:42.866267 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.866284 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 10:48:42 crc kubenswrapper[4679]: E1202 10:48:42.866299 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="registry-server" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.866375 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="registry-server" Dec 02 10:48:42 crc kubenswrapper[4679]: E1202 10:48:42.866394 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="extract-utilities" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.866402 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="extract-utilities" Dec 02 10:48:42 crc kubenswrapper[4679]: E1202 10:48:42.866414 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="extract-content" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.866421 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="extract-content" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.866610 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="082e7b97-c1d2-4ac7-8d37-e36e64bfb0b5" containerName="registry-server" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.866647 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="2119b0d5-f21e-4694-9fc8-7bf3bed24ecc" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.867346 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.869636 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.869713 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.869912 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.878231 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.881029 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv"] Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.923586 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.923858 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw7rd\" (UniqueName: \"kubernetes.io/projected/c6a42827-37a6-46ee-bc6a-925b0434d21a-kube-api-access-rw7rd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:42 crc kubenswrapper[4679]: I1202 10:48:42.923906 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.025976 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.026160 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw7rd\" (UniqueName: \"kubernetes.io/projected/c6a42827-37a6-46ee-bc6a-925b0434d21a-kube-api-access-rw7rd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.026199 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.031412 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.031752 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.044369 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw7rd\" (UniqueName: \"kubernetes.io/projected/c6a42827-37a6-46ee-bc6a-925b0434d21a-kube-api-access-rw7rd\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-thzwv\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.186094 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.720173 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv"] Dec 02 10:48:43 crc kubenswrapper[4679]: I1202 10:48:43.775348 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" event={"ID":"c6a42827-37a6-46ee-bc6a-925b0434d21a","Type":"ContainerStarted","Data":"fd0dd594e8351892d8552dd8351c0bf4e9b7e07e5cec396d2351a861646d82be"} Dec 02 10:48:44 crc kubenswrapper[4679]: I1202 10:48:44.788693 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" event={"ID":"c6a42827-37a6-46ee-bc6a-925b0434d21a","Type":"ContainerStarted","Data":"9ddd10e8cc1538d10e8e9f660a3822a77cbf4108c6b523a0bf1e228ad2b3c1a6"} Dec 02 10:48:45 crc kubenswrapper[4679]: I1202 10:48:45.825956 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" podStartSLOduration=3.261611791 podStartE2EDuration="3.82593837s" podCreationTimestamp="2025-12-02 10:48:42 +0000 UTC" firstStartedPulling="2025-12-02 10:48:43.726721996 +0000 UTC m=+1777.056860866" lastFinishedPulling="2025-12-02 10:48:44.291048575 +0000 UTC m=+1777.621187445" observedRunningTime="2025-12-02 10:48:45.817994121 +0000 UTC m=+1779.148132981" watchObservedRunningTime="2025-12-02 10:48:45.82593837 +0000 UTC m=+1779.156077230" Dec 02 10:48:46 crc kubenswrapper[4679]: I1202 10:48:46.915807 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:48:46 crc kubenswrapper[4679]: E1202 10:48:46.916248 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:48:53 crc kubenswrapper[4679]: I1202 10:48:53.079191 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-zs6tz"] Dec 02 10:48:53 crc kubenswrapper[4679]: I1202 10:48:53.093605 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-zs6tz"] Dec 02 10:48:53 crc kubenswrapper[4679]: I1202 10:48:53.102539 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-w58bz"] Dec 02 10:48:53 crc kubenswrapper[4679]: I1202 10:48:53.110403 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-w58bz"] Dec 02 10:48:54 crc kubenswrapper[4679]: I1202 10:48:54.924341 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d59898e-672f-4699-bf31-a964cd84913b" path="/var/lib/kubelet/pods/8d59898e-672f-4699-bf31-a964cd84913b/volumes" Dec 02 10:48:54 crc kubenswrapper[4679]: I1202 10:48:54.925693 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1284622-943e-4f14-b59c-79eb084e8ea5" path="/var/lib/kubelet/pods/e1284622-943e-4f14-b59c-79eb084e8ea5/volumes" Dec 02 10:48:57 crc kubenswrapper[4679]: I1202 10:48:57.911726 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:48:57 crc kubenswrapper[4679]: E1202 10:48:57.912062 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:49:05 crc kubenswrapper[4679]: I1202 10:49:05.040459 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-9kjcb"] Dec 02 10:49:05 crc kubenswrapper[4679]: I1202 10:49:05.050601 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-86lxg"] Dec 02 10:49:05 crc kubenswrapper[4679]: I1202 10:49:05.073427 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-9kjcb"] Dec 02 10:49:05 crc kubenswrapper[4679]: I1202 10:49:05.092880 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9qtfk"] Dec 02 10:49:05 crc kubenswrapper[4679]: I1202 10:49:05.106760 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-86lxg"] Dec 02 10:49:05 crc kubenswrapper[4679]: I1202 10:49:05.119468 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9qtfk"] Dec 02 10:49:06 crc kubenswrapper[4679]: I1202 10:49:06.922848 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="055921c7-b108-40a9-a4da-2947ca27c2d8" path="/var/lib/kubelet/pods/055921c7-b108-40a9-a4da-2947ca27c2d8/volumes" Dec 02 10:49:06 crc kubenswrapper[4679]: I1202 10:49:06.923543 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b" path="/var/lib/kubelet/pods/6622d06e-f9ba-4b79-b9a3-a52a8b8b6a5b/volumes" Dec 02 10:49:06 crc kubenswrapper[4679]: I1202 10:49:06.924065 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f071f49e-c3e2-4919-a930-dd5fd654cf55" path="/var/lib/kubelet/pods/f071f49e-c3e2-4919-a930-dd5fd654cf55/volumes" Dec 02 10:49:12 crc kubenswrapper[4679]: I1202 10:49:12.909438 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:49:12 crc kubenswrapper[4679]: E1202 10:49:12.910338 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:49:20 crc kubenswrapper[4679]: I1202 10:49:20.058812 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-bqb7p"] Dec 02 10:49:20 crc kubenswrapper[4679]: I1202 10:49:20.067065 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-bqb7p"] Dec 02 10:49:20 crc kubenswrapper[4679]: I1202 10:49:20.924087 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048db06f-54db-4862-b8c8-b7e9cbb55095" path="/var/lib/kubelet/pods/048db06f-54db-4862-b8c8-b7e9cbb55095/volumes" Dec 02 10:49:21 crc kubenswrapper[4679]: I1202 10:49:21.717283 4679 scope.go:117] "RemoveContainer" containerID="16c7e627e2b769fec6fded2a0edc40f16b7e1e272141289d48ed3bc7171d69c4" Dec 02 10:49:21 crc kubenswrapper[4679]: I1202 10:49:21.750406 4679 scope.go:117] "RemoveContainer" containerID="02285a941c9dc5583f689ea9856449177dc4724a9eda53ccf9257492f7eedcce" Dec 02 10:49:21 crc kubenswrapper[4679]: I1202 10:49:21.792773 4679 scope.go:117] "RemoveContainer" containerID="63c81f5df5186090ed9c4c7eeaa433b87cd234a0fd8ab2ba4c543da186a82336" Dec 02 10:49:21 crc kubenswrapper[4679]: I1202 10:49:21.843090 4679 scope.go:117] "RemoveContainer" containerID="50450f1a8db17111a960666ee3f2b7e9231d0580c967b8b19c101e1179dc09a0" Dec 02 10:49:21 crc kubenswrapper[4679]: I1202 10:49:21.898504 4679 scope.go:117] "RemoveContainer" containerID="3dfec170a3ca144e37522da5e343af3fd966f5d6066ed9357992498cfa914b8d" Dec 02 10:49:21 crc kubenswrapper[4679]: I1202 10:49:21.940590 4679 scope.go:117] "RemoveContainer" containerID="4807083e9b2f93e0679d1db73827ffaa4974fc6cc8c70e41542529d1e28db2f8" Dec 02 10:49:28 crc kubenswrapper[4679]: I1202 10:49:28.483496 4679 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-sxgbw" podUID="ebbb7257-823e-49a1-af4c-703b979de700" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 10:49:28 crc kubenswrapper[4679]: I1202 10:49:28.487758 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:49:28 crc kubenswrapper[4679]: E1202 10:49:28.487997 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:49:42 crc kubenswrapper[4679]: I1202 10:49:42.909275 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:49:42 crc kubenswrapper[4679]: E1202 10:49:42.910416 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:49:57 crc kubenswrapper[4679]: I1202 10:49:57.823394 4679 generic.go:334] "Generic (PLEG): container finished" podID="c6a42827-37a6-46ee-bc6a-925b0434d21a" containerID="9ddd10e8cc1538d10e8e9f660a3822a77cbf4108c6b523a0bf1e228ad2b3c1a6" exitCode=0 Dec 02 10:49:57 crc kubenswrapper[4679]: I1202 10:49:57.823477 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" event={"ID":"c6a42827-37a6-46ee-bc6a-925b0434d21a","Type":"ContainerDied","Data":"9ddd10e8cc1538d10e8e9f660a3822a77cbf4108c6b523a0bf1e228ad2b3c1a6"} Dec 02 10:49:57 crc kubenswrapper[4679]: I1202 10:49:57.908938 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:49:57 crc kubenswrapper[4679]: E1202 10:49:57.909198 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.255740 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.393956 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw7rd\" (UniqueName: \"kubernetes.io/projected/c6a42827-37a6-46ee-bc6a-925b0434d21a-kube-api-access-rw7rd\") pod \"c6a42827-37a6-46ee-bc6a-925b0434d21a\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.394475 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-ssh-key\") pod \"c6a42827-37a6-46ee-bc6a-925b0434d21a\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.394633 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory\") pod \"c6a42827-37a6-46ee-bc6a-925b0434d21a\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.399859 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6a42827-37a6-46ee-bc6a-925b0434d21a-kube-api-access-rw7rd" (OuterVolumeSpecName: "kube-api-access-rw7rd") pod "c6a42827-37a6-46ee-bc6a-925b0434d21a" (UID: "c6a42827-37a6-46ee-bc6a-925b0434d21a"). InnerVolumeSpecName "kube-api-access-rw7rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:49:59 crc kubenswrapper[4679]: E1202 10:49:59.422621 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory podName:c6a42827-37a6-46ee-bc6a-925b0434d21a nodeName:}" failed. No retries permitted until 2025-12-02 10:49:59.922592199 +0000 UTC m=+1853.252731059 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory") pod "c6a42827-37a6-46ee-bc6a-925b0434d21a" (UID: "c6a42827-37a6-46ee-bc6a-925b0434d21a") : error deleting /var/lib/kubelet/pods/c6a42827-37a6-46ee-bc6a-925b0434d21a/volume-subpaths: remove /var/lib/kubelet/pods/c6a42827-37a6-46ee-bc6a-925b0434d21a/volume-subpaths: no such file or directory Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.425146 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6a42827-37a6-46ee-bc6a-925b0434d21a" (UID: "c6a42827-37a6-46ee-bc6a-925b0434d21a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.496954 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw7rd\" (UniqueName: \"kubernetes.io/projected/c6a42827-37a6-46ee-bc6a-925b0434d21a-kube-api-access-rw7rd\") on node \"crc\" DevicePath \"\"" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.496990 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.842251 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" event={"ID":"c6a42827-37a6-46ee-bc6a-925b0434d21a","Type":"ContainerDied","Data":"fd0dd594e8351892d8552dd8351c0bf4e9b7e07e5cec396d2351a861646d82be"} Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.842342 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-thzwv" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.842365 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd0dd594e8351892d8552dd8351c0bf4e9b7e07e5cec396d2351a861646d82be" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.956048 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms"] Dec 02 10:49:59 crc kubenswrapper[4679]: E1202 10:49:59.956666 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a42827-37a6-46ee-bc6a-925b0434d21a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.956693 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a42827-37a6-46ee-bc6a-925b0434d21a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.956902 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a42827-37a6-46ee-bc6a-925b0434d21a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.957689 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:49:59 crc kubenswrapper[4679]: I1202 10:49:59.965706 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms"] Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.007283 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory\") pod \"c6a42827-37a6-46ee-bc6a-925b0434d21a\" (UID: \"c6a42827-37a6-46ee-bc6a-925b0434d21a\") " Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.011495 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory" (OuterVolumeSpecName: "inventory") pod "c6a42827-37a6-46ee-bc6a-925b0434d21a" (UID: "c6a42827-37a6-46ee-bc6a-925b0434d21a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.109834 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.109896 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c44ln\" (UniqueName: \"kubernetes.io/projected/1420f674-925e-449b-8e18-0e97225f3958-kube-api-access-c44ln\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.110030 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.110886 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6a42827-37a6-46ee-bc6a-925b0434d21a-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.213046 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.213125 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c44ln\" (UniqueName: \"kubernetes.io/projected/1420f674-925e-449b-8e18-0e97225f3958-kube-api-access-c44ln\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.213241 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.217421 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.217838 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.229912 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c44ln\" (UniqueName: \"kubernetes.io/projected/1420f674-925e-449b-8e18-0e97225f3958-kube-api-access-c44ln\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5nrms\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.279083 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.802677 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms"] Dec 02 10:50:00 crc kubenswrapper[4679]: I1202 10:50:00.851030 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" event={"ID":"1420f674-925e-449b-8e18-0e97225f3958","Type":"ContainerStarted","Data":"3c7a12d6530cf1d6bbb0afee7e6cadbaf57f99be75da7df880283013254d3f5c"} Dec 02 10:50:02 crc kubenswrapper[4679]: I1202 10:50:02.870059 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" event={"ID":"1420f674-925e-449b-8e18-0e97225f3958","Type":"ContainerStarted","Data":"bb2b3d0ffc02167b72c0f4a536d79cce0e9665cb0a3d80e25cb4454fa7db14f3"} Dec 02 10:50:02 crc kubenswrapper[4679]: I1202 10:50:02.893229 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" podStartSLOduration=2.9907345850000002 podStartE2EDuration="3.893211709s" podCreationTimestamp="2025-12-02 10:49:59 +0000 UTC" firstStartedPulling="2025-12-02 10:50:00.804496419 +0000 UTC m=+1854.134635279" lastFinishedPulling="2025-12-02 10:50:01.706973543 +0000 UTC m=+1855.037112403" observedRunningTime="2025-12-02 10:50:02.888833443 +0000 UTC m=+1856.218972313" watchObservedRunningTime="2025-12-02 10:50:02.893211709 +0000 UTC m=+1856.223350569" Dec 02 10:50:06 crc kubenswrapper[4679]: I1202 10:50:06.907919 4679 generic.go:334] "Generic (PLEG): container finished" podID="1420f674-925e-449b-8e18-0e97225f3958" containerID="bb2b3d0ffc02167b72c0f4a536d79cce0e9665cb0a3d80e25cb4454fa7db14f3" exitCode=0 Dec 02 10:50:06 crc kubenswrapper[4679]: I1202 10:50:06.933391 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" event={"ID":"1420f674-925e-449b-8e18-0e97225f3958","Type":"ContainerDied","Data":"bb2b3d0ffc02167b72c0f4a536d79cce0e9665cb0a3d80e25cb4454fa7db14f3"} Dec 02 10:50:07 crc kubenswrapper[4679]: I1202 10:50:07.051383 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lvb5d"] Dec 02 10:50:07 crc kubenswrapper[4679]: I1202 10:50:07.059939 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-dbb7-account-create-update-mc5lj"] Dec 02 10:50:07 crc kubenswrapper[4679]: I1202 10:50:07.067683 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-fj9wm"] Dec 02 10:50:07 crc kubenswrapper[4679]: I1202 10:50:07.076514 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-fj9wm"] Dec 02 10:50:07 crc kubenswrapper[4679]: I1202 10:50:07.084616 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lvb5d"] Dec 02 10:50:07 crc kubenswrapper[4679]: I1202 10:50:07.092748 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-dbb7-account-create-update-mc5lj"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.030339 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b5tpv"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.038544 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4f48-account-create-update-5kwkr"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.048251 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-feeb-account-create-update-v5pzb"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.055535 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b5tpv"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.066484 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-feeb-account-create-update-v5pzb"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.071708 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4f48-account-create-update-5kwkr"] Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.336281 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.468922 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c44ln\" (UniqueName: \"kubernetes.io/projected/1420f674-925e-449b-8e18-0e97225f3958-kube-api-access-c44ln\") pod \"1420f674-925e-449b-8e18-0e97225f3958\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.469337 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-inventory\") pod \"1420f674-925e-449b-8e18-0e97225f3958\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.469455 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-ssh-key\") pod \"1420f674-925e-449b-8e18-0e97225f3958\" (UID: \"1420f674-925e-449b-8e18-0e97225f3958\") " Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.474997 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1420f674-925e-449b-8e18-0e97225f3958-kube-api-access-c44ln" (OuterVolumeSpecName: "kube-api-access-c44ln") pod "1420f674-925e-449b-8e18-0e97225f3958" (UID: "1420f674-925e-449b-8e18-0e97225f3958"). InnerVolumeSpecName "kube-api-access-c44ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.497992 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1420f674-925e-449b-8e18-0e97225f3958" (UID: "1420f674-925e-449b-8e18-0e97225f3958"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.509256 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-inventory" (OuterVolumeSpecName: "inventory") pod "1420f674-925e-449b-8e18-0e97225f3958" (UID: "1420f674-925e-449b-8e18-0e97225f3958"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.571999 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c44ln\" (UniqueName: \"kubernetes.io/projected/1420f674-925e-449b-8e18-0e97225f3958-kube-api-access-c44ln\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.572029 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.572039 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1420f674-925e-449b-8e18-0e97225f3958-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.932703 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d1a8e7c-e69d-4400-84c2-27537b0593f0" path="/var/lib/kubelet/pods/0d1a8e7c-e69d-4400-84c2-27537b0593f0/volumes" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.933256 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30145675-4813-4416-a06e-1487bf15c0fe" path="/var/lib/kubelet/pods/30145675-4813-4416-a06e-1487bf15c0fe/volumes" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.934076 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc64597-c029-4073-8f2f-9d25583c1767" path="/var/lib/kubelet/pods/5bc64597-c029-4073-8f2f-9d25583c1767/volumes" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.934715 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3" path="/var/lib/kubelet/pods/89bd0ea1-eeff-4b25-9aa8-06125cd4c0d3/volumes" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.934880 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.935791 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afead6b9-e07e-4d06-af9a-2f876bb4c5e8" path="/var/lib/kubelet/pods/afead6b9-e07e-4d06-af9a-2f876bb4c5e8/volumes" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.936437 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c372bf97-ea32-4ca7-99cc-9ead0f8e277f" path="/var/lib/kubelet/pods/c372bf97-ea32-4ca7-99cc-9ead0f8e277f/volumes" Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.937232 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5nrms" event={"ID":"1420f674-925e-449b-8e18-0e97225f3958","Type":"ContainerDied","Data":"3c7a12d6530cf1d6bbb0afee7e6cadbaf57f99be75da7df880283013254d3f5c"} Dec 02 10:50:08 crc kubenswrapper[4679]: I1202 10:50:08.937334 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c7a12d6530cf1d6bbb0afee7e6cadbaf57f99be75da7df880283013254d3f5c" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.017679 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr"] Dec 02 10:50:09 crc kubenswrapper[4679]: E1202 10:50:09.018070 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1420f674-925e-449b-8e18-0e97225f3958" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.018088 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1420f674-925e-449b-8e18-0e97225f3958" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.018273 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="1420f674-925e-449b-8e18-0e97225f3958" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.019183 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.021704 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.021807 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.024702 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.024891 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.044325 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr"] Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.082629 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.082904 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.083219 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdmf8\" (UniqueName: \"kubernetes.io/projected/645181f0-3043-4bad-908e-8d57baa1d9b9-kube-api-access-cdmf8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.184891 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.185053 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmf8\" (UniqueName: \"kubernetes.io/projected/645181f0-3043-4bad-908e-8d57baa1d9b9-kube-api-access-cdmf8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.185121 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.188746 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.189733 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.203585 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdmf8\" (UniqueName: \"kubernetes.io/projected/645181f0-3043-4bad-908e-8d57baa1d9b9-kube-api-access-cdmf8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-nr9dr\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.345094 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.832460 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr"] Dec 02 10:50:09 crc kubenswrapper[4679]: I1202 10:50:09.943553 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" event={"ID":"645181f0-3043-4bad-908e-8d57baa1d9b9","Type":"ContainerStarted","Data":"ebad0337d909abd03bd2a4c36dd20beac05304ccd480826d202c122c6610c342"} Dec 02 10:50:10 crc kubenswrapper[4679]: I1202 10:50:10.956262 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" event={"ID":"645181f0-3043-4bad-908e-8d57baa1d9b9","Type":"ContainerStarted","Data":"c4b90677267ee47efdeef564aad9b0c321f206af52c944db638b84c390ca94da"} Dec 02 10:50:10 crc kubenswrapper[4679]: I1202 10:50:10.978724 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" podStartSLOduration=2.429491374 podStartE2EDuration="2.978708568s" podCreationTimestamp="2025-12-02 10:50:08 +0000 UTC" firstStartedPulling="2025-12-02 10:50:09.838905919 +0000 UTC m=+1863.169044779" lastFinishedPulling="2025-12-02 10:50:10.388123113 +0000 UTC m=+1863.718261973" observedRunningTime="2025-12-02 10:50:10.975391072 +0000 UTC m=+1864.305529942" watchObservedRunningTime="2025-12-02 10:50:10.978708568 +0000 UTC m=+1864.308847428" Dec 02 10:50:12 crc kubenswrapper[4679]: I1202 10:50:12.909218 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:50:12 crc kubenswrapper[4679]: E1202 10:50:12.909695 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:50:22 crc kubenswrapper[4679]: I1202 10:50:22.086681 4679 scope.go:117] "RemoveContainer" containerID="c9dc10bfade2c586c041c88061a89a592c6213c7d802b2e701d808de41573cd9" Dec 02 10:50:22 crc kubenswrapper[4679]: I1202 10:50:22.117730 4679 scope.go:117] "RemoveContainer" containerID="aed740d8d2ec18baeaea3f61355f29c1cbcedf2576bafb2b36c24b5aef0f4c07" Dec 02 10:50:22 crc kubenswrapper[4679]: I1202 10:50:22.170530 4679 scope.go:117] "RemoveContainer" containerID="a8014e75c6e0af69dda28c0fe97bc90313f7c37f211b5af08a20400efabcb901" Dec 02 10:50:22 crc kubenswrapper[4679]: I1202 10:50:22.210231 4679 scope.go:117] "RemoveContainer" containerID="2348b19a4bc333500eeda889787e776da3d58468521f673fa9a305dd77b96998" Dec 02 10:50:22 crc kubenswrapper[4679]: I1202 10:50:22.254740 4679 scope.go:117] "RemoveContainer" containerID="4731d335898e37ad10b75005bac9e8b7df4021a8932215936289081b270620ee" Dec 02 10:50:22 crc kubenswrapper[4679]: I1202 10:50:22.307251 4679 scope.go:117] "RemoveContainer" containerID="d3bd56934222bcfb7262c2adcfaa69189e6ad959c857cf212711ec2ee2a820ff" Dec 02 10:50:23 crc kubenswrapper[4679]: I1202 10:50:23.908701 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:50:23 crc kubenswrapper[4679]: E1202 10:50:23.910338 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:50:33 crc kubenswrapper[4679]: I1202 10:50:33.040927 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bm8fd"] Dec 02 10:50:33 crc kubenswrapper[4679]: I1202 10:50:33.049342 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bm8fd"] Dec 02 10:50:34 crc kubenswrapper[4679]: I1202 10:50:34.919834 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b287fe6f-16df-4020-85b8-21ebdc224103" path="/var/lib/kubelet/pods/b287fe6f-16df-4020-85b8-21ebdc224103/volumes" Dec 02 10:50:35 crc kubenswrapper[4679]: I1202 10:50:35.908932 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:50:35 crc kubenswrapper[4679]: E1202 10:50:35.909379 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:50:46 crc kubenswrapper[4679]: I1202 10:50:46.923391 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:50:46 crc kubenswrapper[4679]: E1202 10:50:46.924448 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:50:52 crc kubenswrapper[4679]: E1202 10:50:52.685147 4679 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod645181f0_3043_4bad_908e_8d57baa1d9b9.slice/crio-conmon-c4b90677267ee47efdeef564aad9b0c321f206af52c944db638b84c390ca94da.scope\": RecentStats: unable to find data in memory cache]" Dec 02 10:50:53 crc kubenswrapper[4679]: I1202 10:50:53.319789 4679 generic.go:334] "Generic (PLEG): container finished" podID="645181f0-3043-4bad-908e-8d57baa1d9b9" containerID="c4b90677267ee47efdeef564aad9b0c321f206af52c944db638b84c390ca94da" exitCode=0 Dec 02 10:50:53 crc kubenswrapper[4679]: I1202 10:50:53.319860 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" event={"ID":"645181f0-3043-4bad-908e-8d57baa1d9b9","Type":"ContainerDied","Data":"c4b90677267ee47efdeef564aad9b0c321f206af52c944db638b84c390ca94da"} Dec 02 10:50:54 crc kubenswrapper[4679]: I1202 10:50:54.888406 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.056271 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-inventory\") pod \"645181f0-3043-4bad-908e-8d57baa1d9b9\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.056451 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-ssh-key\") pod \"645181f0-3043-4bad-908e-8d57baa1d9b9\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.056699 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdmf8\" (UniqueName: \"kubernetes.io/projected/645181f0-3043-4bad-908e-8d57baa1d9b9-kube-api-access-cdmf8\") pod \"645181f0-3043-4bad-908e-8d57baa1d9b9\" (UID: \"645181f0-3043-4bad-908e-8d57baa1d9b9\") " Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.065576 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645181f0-3043-4bad-908e-8d57baa1d9b9-kube-api-access-cdmf8" (OuterVolumeSpecName: "kube-api-access-cdmf8") pod "645181f0-3043-4bad-908e-8d57baa1d9b9" (UID: "645181f0-3043-4bad-908e-8d57baa1d9b9"). InnerVolumeSpecName "kube-api-access-cdmf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.086347 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "645181f0-3043-4bad-908e-8d57baa1d9b9" (UID: "645181f0-3043-4bad-908e-8d57baa1d9b9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.107185 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-inventory" (OuterVolumeSpecName: "inventory") pod "645181f0-3043-4bad-908e-8d57baa1d9b9" (UID: "645181f0-3043-4bad-908e-8d57baa1d9b9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.159877 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.159903 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645181f0-3043-4bad-908e-8d57baa1d9b9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.159912 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdmf8\" (UniqueName: \"kubernetes.io/projected/645181f0-3043-4bad-908e-8d57baa1d9b9-kube-api-access-cdmf8\") on node \"crc\" DevicePath \"\"" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.340551 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" event={"ID":"645181f0-3043-4bad-908e-8d57baa1d9b9","Type":"ContainerDied","Data":"ebad0337d909abd03bd2a4c36dd20beac05304ccd480826d202c122c6610c342"} Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.340596 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebad0337d909abd03bd2a4c36dd20beac05304ccd480826d202c122c6610c342" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.340641 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-nr9dr" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.421689 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m"] Dec 02 10:50:55 crc kubenswrapper[4679]: E1202 10:50:55.422091 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645181f0-3043-4bad-908e-8d57baa1d9b9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.422114 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="645181f0-3043-4bad-908e-8d57baa1d9b9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.422417 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="645181f0-3043-4bad-908e-8d57baa1d9b9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.423083 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.428659 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.428802 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.428900 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.429236 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.433692 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m"] Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.568483 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvfbw\" (UniqueName: \"kubernetes.io/projected/9947dc7b-205c-463b-98d9-07759646f56f-kube-api-access-zvfbw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.568558 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.568704 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.670833 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.670918 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvfbw\" (UniqueName: \"kubernetes.io/projected/9947dc7b-205c-463b-98d9-07759646f56f-kube-api-access-zvfbw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.670956 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.676156 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.676225 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.708536 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvfbw\" (UniqueName: \"kubernetes.io/projected/9947dc7b-205c-463b-98d9-07759646f56f-kube-api-access-zvfbw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ff56m\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:55 crc kubenswrapper[4679]: I1202 10:50:55.746673 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.059527 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-md655"] Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.069662 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ttsxh"] Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.080502 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-md655"] Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.089501 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ttsxh"] Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.259641 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m"] Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.348799 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" event={"ID":"9947dc7b-205c-463b-98d9-07759646f56f","Type":"ContainerStarted","Data":"7107ebd01855ace6d6cbdf4459f4efbccbfd7892d9f18ef60473aed4f76fd9ef"} Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.925150 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7abb7421-f56a-462b-baea-b0030a020bd5" path="/var/lib/kubelet/pods/7abb7421-f56a-462b-baea-b0030a020bd5/volumes" Dec 02 10:50:56 crc kubenswrapper[4679]: I1202 10:50:56.925724 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c5b0d9f-8c3f-478e-9622-44513c5bd88c" path="/var/lib/kubelet/pods/7c5b0d9f-8c3f-478e-9622-44513c5bd88c/volumes" Dec 02 10:50:58 crc kubenswrapper[4679]: I1202 10:50:58.368365 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" event={"ID":"9947dc7b-205c-463b-98d9-07759646f56f","Type":"ContainerStarted","Data":"71dc3e2f0ecf6c0dbb0fa7d9f22acb044f702fca07f83b2bd3a537c0f022b2b0"} Dec 02 10:50:59 crc kubenswrapper[4679]: I1202 10:50:59.398249 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" podStartSLOduration=2.732436006 podStartE2EDuration="4.398218549s" podCreationTimestamp="2025-12-02 10:50:55 +0000 UTC" firstStartedPulling="2025-12-02 10:50:56.272924933 +0000 UTC m=+1909.603063783" lastFinishedPulling="2025-12-02 10:50:57.938707456 +0000 UTC m=+1911.268846326" observedRunningTime="2025-12-02 10:50:59.398093146 +0000 UTC m=+1912.728232036" watchObservedRunningTime="2025-12-02 10:50:59.398218549 +0000 UTC m=+1912.728357449" Dec 02 10:50:59 crc kubenswrapper[4679]: I1202 10:50:59.909504 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:50:59 crc kubenswrapper[4679]: E1202 10:50:59.909830 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:51:14 crc kubenswrapper[4679]: I1202 10:51:14.910407 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:51:14 crc kubenswrapper[4679]: E1202 10:51:14.911758 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:51:22 crc kubenswrapper[4679]: I1202 10:51:22.448870 4679 scope.go:117] "RemoveContainer" containerID="184efdb137bd8381f06c71a3d4fc9e363a201b001cbb34d0699471c10b2dcab2" Dec 02 10:51:22 crc kubenswrapper[4679]: I1202 10:51:22.497898 4679 scope.go:117] "RemoveContainer" containerID="a8bd6dae67bddf4f4f5d13f500e2c7e7950a36902cbcb8e667aa14132735a428" Dec 02 10:51:22 crc kubenswrapper[4679]: I1202 10:51:22.558230 4679 scope.go:117] "RemoveContainer" containerID="a053164b30b3974286b3971af458f15d99bfbffdcfe0513f8395aa141c0e7d26" Dec 02 10:51:29 crc kubenswrapper[4679]: I1202 10:51:29.909731 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:51:29 crc kubenswrapper[4679]: E1202 10:51:29.910585 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:51:40 crc kubenswrapper[4679]: I1202 10:51:40.041526 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gg8gl"] Dec 02 10:51:40 crc kubenswrapper[4679]: I1202 10:51:40.054478 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gg8gl"] Dec 02 10:51:40 crc kubenswrapper[4679]: I1202 10:51:40.918289 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4804674-9946-4546-aca6-f12ca930cbf8" path="/var/lib/kubelet/pods/d4804674-9946-4546-aca6-f12ca930cbf8/volumes" Dec 02 10:51:44 crc kubenswrapper[4679]: I1202 10:51:44.909330 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:51:44 crc kubenswrapper[4679]: E1202 10:51:44.910156 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:51:48 crc kubenswrapper[4679]: I1202 10:51:48.837203 4679 generic.go:334] "Generic (PLEG): container finished" podID="9947dc7b-205c-463b-98d9-07759646f56f" containerID="71dc3e2f0ecf6c0dbb0fa7d9f22acb044f702fca07f83b2bd3a537c0f022b2b0" exitCode=0 Dec 02 10:51:48 crc kubenswrapper[4679]: I1202 10:51:48.837322 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" event={"ID":"9947dc7b-205c-463b-98d9-07759646f56f","Type":"ContainerDied","Data":"71dc3e2f0ecf6c0dbb0fa7d9f22acb044f702fca07f83b2bd3a537c0f022b2b0"} Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.251178 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.389870 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-inventory\") pod \"9947dc7b-205c-463b-98d9-07759646f56f\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.389972 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key\") pod \"9947dc7b-205c-463b-98d9-07759646f56f\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.390110 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvfbw\" (UniqueName: \"kubernetes.io/projected/9947dc7b-205c-463b-98d9-07759646f56f-kube-api-access-zvfbw\") pod \"9947dc7b-205c-463b-98d9-07759646f56f\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.397006 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9947dc7b-205c-463b-98d9-07759646f56f-kube-api-access-zvfbw" (OuterVolumeSpecName: "kube-api-access-zvfbw") pod "9947dc7b-205c-463b-98d9-07759646f56f" (UID: "9947dc7b-205c-463b-98d9-07759646f56f"). InnerVolumeSpecName "kube-api-access-zvfbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:51:50 crc kubenswrapper[4679]: E1202 10:51:50.427216 4679 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key podName:9947dc7b-205c-463b-98d9-07759646f56f nodeName:}" failed. No retries permitted until 2025-12-02 10:51:50.927185985 +0000 UTC m=+1964.257324845 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key") pod "9947dc7b-205c-463b-98d9-07759646f56f" (UID: "9947dc7b-205c-463b-98d9-07759646f56f") : error deleting /var/lib/kubelet/pods/9947dc7b-205c-463b-98d9-07759646f56f/volume-subpaths: remove /var/lib/kubelet/pods/9947dc7b-205c-463b-98d9-07759646f56f/volume-subpaths: no such file or directory Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.431067 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-inventory" (OuterVolumeSpecName: "inventory") pod "9947dc7b-205c-463b-98d9-07759646f56f" (UID: "9947dc7b-205c-463b-98d9-07759646f56f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.492207 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvfbw\" (UniqueName: \"kubernetes.io/projected/9947dc7b-205c-463b-98d9-07759646f56f-kube-api-access-zvfbw\") on node \"crc\" DevicePath \"\"" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.492241 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.856765 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" event={"ID":"9947dc7b-205c-463b-98d9-07759646f56f","Type":"ContainerDied","Data":"7107ebd01855ace6d6cbdf4459f4efbccbfd7892d9f18ef60473aed4f76fd9ef"} Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.856815 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7107ebd01855ace6d6cbdf4459f4efbccbfd7892d9f18ef60473aed4f76fd9ef" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.856819 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ff56m" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.969349 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jtmsz"] Dec 02 10:51:50 crc kubenswrapper[4679]: E1202 10:51:50.970144 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9947dc7b-205c-463b-98d9-07759646f56f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.970171 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="9947dc7b-205c-463b-98d9-07759646f56f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.970450 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="9947dc7b-205c-463b-98d9-07759646f56f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.971189 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:50 crc kubenswrapper[4679]: I1202 10:51:50.985531 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jtmsz"] Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.001112 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key\") pod \"9947dc7b-205c-463b-98d9-07759646f56f\" (UID: \"9947dc7b-205c-463b-98d9-07759646f56f\") " Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.008491 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9947dc7b-205c-463b-98d9-07759646f56f" (UID: "9947dc7b-205c-463b-98d9-07759646f56f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.103397 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.103520 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.103537 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7lkc\" (UniqueName: \"kubernetes.io/projected/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-kube-api-access-b7lkc\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.103605 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9947dc7b-205c-463b-98d9-07759646f56f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.204976 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.205041 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7lkc\" (UniqueName: \"kubernetes.io/projected/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-kube-api-access-b7lkc\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.205149 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.208426 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.211499 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.225269 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7lkc\" (UniqueName: \"kubernetes.io/projected/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-kube-api-access-b7lkc\") pod \"ssh-known-hosts-edpm-deployment-jtmsz\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.378249 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.967778 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jtmsz"] Dec 02 10:51:51 crc kubenswrapper[4679]: I1202 10:51:51.970001 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:51:52 crc kubenswrapper[4679]: I1202 10:51:52.877044 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" event={"ID":"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7","Type":"ContainerStarted","Data":"746046d644ae062550159f33d2a2ac33b09b323546eaca41ce47cb5035bc5355"} Dec 02 10:51:55 crc kubenswrapper[4679]: I1202 10:51:55.906968 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" event={"ID":"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7","Type":"ContainerStarted","Data":"6886106d51cc86f0079847c0b7df5a3bf2da1c0821d7930955b89d5de7f19607"} Dec 02 10:51:55 crc kubenswrapper[4679]: I1202 10:51:55.924405 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" podStartSLOduration=3.176751164 podStartE2EDuration="5.924390186s" podCreationTimestamp="2025-12-02 10:51:50 +0000 UTC" firstStartedPulling="2025-12-02 10:51:51.969714543 +0000 UTC m=+1965.299853403" lastFinishedPulling="2025-12-02 10:51:54.717353565 +0000 UTC m=+1968.047492425" observedRunningTime="2025-12-02 10:51:55.924230251 +0000 UTC m=+1969.254369111" watchObservedRunningTime="2025-12-02 10:51:55.924390186 +0000 UTC m=+1969.254529046" Dec 02 10:51:58 crc kubenswrapper[4679]: I1202 10:51:58.908647 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:51:58 crc kubenswrapper[4679]: E1202 10:51:58.909199 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:52:01 crc kubenswrapper[4679]: I1202 10:52:01.957867 4679 generic.go:334] "Generic (PLEG): container finished" podID="8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" containerID="6886106d51cc86f0079847c0b7df5a3bf2da1c0821d7930955b89d5de7f19607" exitCode=0 Dec 02 10:52:01 crc kubenswrapper[4679]: I1202 10:52:01.957992 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" event={"ID":"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7","Type":"ContainerDied","Data":"6886106d51cc86f0079847c0b7df5a3bf2da1c0821d7930955b89d5de7f19607"} Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.357107 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.545681 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-ssh-key-openstack-edpm-ipam\") pod \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.545758 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-inventory-0\") pod \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.545899 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7lkc\" (UniqueName: \"kubernetes.io/projected/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-kube-api-access-b7lkc\") pod \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\" (UID: \"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7\") " Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.557609 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-kube-api-access-b7lkc" (OuterVolumeSpecName: "kube-api-access-b7lkc") pod "8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" (UID: "8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7"). InnerVolumeSpecName "kube-api-access-b7lkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.573906 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" (UID: "8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.587612 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" (UID: "8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.648478 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7lkc\" (UniqueName: \"kubernetes.io/projected/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-kube-api-access-b7lkc\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.648712 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.648772 4679 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.976509 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" event={"ID":"8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7","Type":"ContainerDied","Data":"746046d644ae062550159f33d2a2ac33b09b323546eaca41ce47cb5035bc5355"} Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.976548 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="746046d644ae062550159f33d2a2ac33b09b323546eaca41ce47cb5035bc5355" Dec 02 10:52:03 crc kubenswrapper[4679]: I1202 10:52:03.976598 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jtmsz" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.060578 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf"] Dec 02 10:52:04 crc kubenswrapper[4679]: E1202 10:52:04.061080 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" containerName="ssh-known-hosts-edpm-deployment" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.061107 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" containerName="ssh-known-hosts-edpm-deployment" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.061370 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7" containerName="ssh-known-hosts-edpm-deployment" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.062341 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.064370 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.064435 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.064496 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.064782 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.077567 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf"] Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.157505 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.157567 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.157671 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrjfq\" (UniqueName: \"kubernetes.io/projected/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-kube-api-access-zrjfq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.259449 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.259518 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.259571 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrjfq\" (UniqueName: \"kubernetes.io/projected/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-kube-api-access-zrjfq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.265505 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.271992 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.281895 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrjfq\" (UniqueName: \"kubernetes.io/projected/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-kube-api-access-zrjfq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7dlsf\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.389165 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.935545 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf"] Dec 02 10:52:04 crc kubenswrapper[4679]: I1202 10:52:04.986365 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" event={"ID":"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c","Type":"ContainerStarted","Data":"2c9b8adb5993aa2ce4671e999a91d92cc5cb463abc77904ae30efd0392b84b78"} Dec 02 10:52:05 crc kubenswrapper[4679]: I1202 10:52:05.996832 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" event={"ID":"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c","Type":"ContainerStarted","Data":"56753b4183cf6cf625108081663d41cb76b79527581482b2c7b1dcb8d2211638"} Dec 02 10:52:06 crc kubenswrapper[4679]: I1202 10:52:06.019164 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" podStartSLOduration=1.543710388 podStartE2EDuration="2.019145157s" podCreationTimestamp="2025-12-02 10:52:04 +0000 UTC" firstStartedPulling="2025-12-02 10:52:04.946588091 +0000 UTC m=+1978.276726951" lastFinishedPulling="2025-12-02 10:52:05.42202286 +0000 UTC m=+1978.752161720" observedRunningTime="2025-12-02 10:52:06.011206099 +0000 UTC m=+1979.341344969" watchObservedRunningTime="2025-12-02 10:52:06.019145157 +0000 UTC m=+1979.349284017" Dec 02 10:52:10 crc kubenswrapper[4679]: I1202 10:52:10.909264 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:52:10 crc kubenswrapper[4679]: E1202 10:52:10.912207 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:52:14 crc kubenswrapper[4679]: I1202 10:52:14.081070 4679 generic.go:334] "Generic (PLEG): container finished" podID="2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" containerID="56753b4183cf6cf625108081663d41cb76b79527581482b2c7b1dcb8d2211638" exitCode=0 Dec 02 10:52:14 crc kubenswrapper[4679]: I1202 10:52:14.081154 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" event={"ID":"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c","Type":"ContainerDied","Data":"56753b4183cf6cf625108081663d41cb76b79527581482b2c7b1dcb8d2211638"} Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.584738 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.677391 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-inventory\") pod \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.677491 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-ssh-key\") pod \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.677638 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrjfq\" (UniqueName: \"kubernetes.io/projected/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-kube-api-access-zrjfq\") pod \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\" (UID: \"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c\") " Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.683576 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-kube-api-access-zrjfq" (OuterVolumeSpecName: "kube-api-access-zrjfq") pod "2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" (UID: "2fd59ebd-72da-4ad7-93b5-7f51f3a3304c"). InnerVolumeSpecName "kube-api-access-zrjfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.714661 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-inventory" (OuterVolumeSpecName: "inventory") pod "2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" (UID: "2fd59ebd-72da-4ad7-93b5-7f51f3a3304c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.722417 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" (UID: "2fd59ebd-72da-4ad7-93b5-7f51f3a3304c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.780102 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrjfq\" (UniqueName: \"kubernetes.io/projected/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-kube-api-access-zrjfq\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.780535 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:15 crc kubenswrapper[4679]: I1202 10:52:15.780556 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd59ebd-72da-4ad7-93b5-7f51f3a3304c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.104993 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" event={"ID":"2fd59ebd-72da-4ad7-93b5-7f51f3a3304c","Type":"ContainerDied","Data":"2c9b8adb5993aa2ce4671e999a91d92cc5cb463abc77904ae30efd0392b84b78"} Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.105064 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c9b8adb5993aa2ce4671e999a91d92cc5cb463abc77904ae30efd0392b84b78" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.105149 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7dlsf" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.168283 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj"] Dec 02 10:52:16 crc kubenswrapper[4679]: E1202 10:52:16.171058 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.171091 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.171405 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd59ebd-72da-4ad7-93b5-7f51f3a3304c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.172560 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.181947 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.182100 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.182291 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.188530 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj"] Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.191262 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.298592 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.298672 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89c4v\" (UniqueName: \"kubernetes.io/projected/0430fbc5-3d73-4885-8670-61dd256714be-kube-api-access-89c4v\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.298728 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.400445 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.400534 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89c4v\" (UniqueName: \"kubernetes.io/projected/0430fbc5-3d73-4885-8670-61dd256714be-kube-api-access-89c4v\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.400577 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.406532 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.406532 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.418951 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89c4v\" (UniqueName: \"kubernetes.io/projected/0430fbc5-3d73-4885-8670-61dd256714be-kube-api-access-89c4v\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.491253 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:16 crc kubenswrapper[4679]: I1202 10:52:16.843998 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj"] Dec 02 10:52:17 crc kubenswrapper[4679]: I1202 10:52:17.113776 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" event={"ID":"0430fbc5-3d73-4885-8670-61dd256714be","Type":"ContainerStarted","Data":"f361c7f51064b41340e7a3f4e22e029585053c28a9d402e33e6ea496a817ff39"} Dec 02 10:52:18 crc kubenswrapper[4679]: I1202 10:52:18.144653 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" event={"ID":"0430fbc5-3d73-4885-8670-61dd256714be","Type":"ContainerStarted","Data":"435c2586a5f1a9e9e51c1b056c2883953d2f0d815b0c7180f2bac39c37c9ba66"} Dec 02 10:52:18 crc kubenswrapper[4679]: I1202 10:52:18.161647 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" podStartSLOduration=1.673531456 podStartE2EDuration="2.161630097s" podCreationTimestamp="2025-12-02 10:52:16 +0000 UTC" firstStartedPulling="2025-12-02 10:52:16.843268246 +0000 UTC m=+1990.173407106" lastFinishedPulling="2025-12-02 10:52:17.331366887 +0000 UTC m=+1990.661505747" observedRunningTime="2025-12-02 10:52:18.158749874 +0000 UTC m=+1991.488888754" watchObservedRunningTime="2025-12-02 10:52:18.161630097 +0000 UTC m=+1991.491768957" Dec 02 10:52:22 crc kubenswrapper[4679]: I1202 10:52:22.663200 4679 scope.go:117] "RemoveContainer" containerID="375d036778fc2db2dc4e0c6120ae42fe941461104c98914afa3117759d602ce8" Dec 02 10:52:22 crc kubenswrapper[4679]: I1202 10:52:22.910204 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:52:23 crc kubenswrapper[4679]: I1202 10:52:23.188276 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"9c8b1bf853618b96cd693f0b2fd8cdfb8559d98bac63a4c138f60ed0d408a7d0"} Dec 02 10:52:28 crc kubenswrapper[4679]: I1202 10:52:28.234430 4679 generic.go:334] "Generic (PLEG): container finished" podID="0430fbc5-3d73-4885-8670-61dd256714be" containerID="435c2586a5f1a9e9e51c1b056c2883953d2f0d815b0c7180f2bac39c37c9ba66" exitCode=0 Dec 02 10:52:28 crc kubenswrapper[4679]: I1202 10:52:28.234533 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" event={"ID":"0430fbc5-3d73-4885-8670-61dd256714be","Type":"ContainerDied","Data":"435c2586a5f1a9e9e51c1b056c2883953d2f0d815b0c7180f2bac39c37c9ba66"} Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.674061 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.760956 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-inventory\") pod \"0430fbc5-3d73-4885-8670-61dd256714be\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.761062 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-ssh-key\") pod \"0430fbc5-3d73-4885-8670-61dd256714be\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.761265 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89c4v\" (UniqueName: \"kubernetes.io/projected/0430fbc5-3d73-4885-8670-61dd256714be-kube-api-access-89c4v\") pod \"0430fbc5-3d73-4885-8670-61dd256714be\" (UID: \"0430fbc5-3d73-4885-8670-61dd256714be\") " Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.767575 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0430fbc5-3d73-4885-8670-61dd256714be-kube-api-access-89c4v" (OuterVolumeSpecName: "kube-api-access-89c4v") pod "0430fbc5-3d73-4885-8670-61dd256714be" (UID: "0430fbc5-3d73-4885-8670-61dd256714be"). InnerVolumeSpecName "kube-api-access-89c4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.794023 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-inventory" (OuterVolumeSpecName: "inventory") pod "0430fbc5-3d73-4885-8670-61dd256714be" (UID: "0430fbc5-3d73-4885-8670-61dd256714be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.799100 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0430fbc5-3d73-4885-8670-61dd256714be" (UID: "0430fbc5-3d73-4885-8670-61dd256714be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.865835 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.866632 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0430fbc5-3d73-4885-8670-61dd256714be-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.866652 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89c4v\" (UniqueName: \"kubernetes.io/projected/0430fbc5-3d73-4885-8670-61dd256714be-kube-api-access-89c4v\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.875737 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f5jzk"] Dec 02 10:52:29 crc kubenswrapper[4679]: E1202 10:52:29.876350 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0430fbc5-3d73-4885-8670-61dd256714be" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.876369 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0430fbc5-3d73-4885-8670-61dd256714be" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.876620 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0430fbc5-3d73-4885-8670-61dd256714be" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.885398 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.910958 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5jzk"] Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.968760 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vmhm\" (UniqueName: \"kubernetes.io/projected/08df06a8-008b-408b-b2cb-209f5cd2a5e5-kube-api-access-2vmhm\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.968834 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-utilities\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:29 crc kubenswrapper[4679]: I1202 10:52:29.969059 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-catalog-content\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.071412 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vmhm\" (UniqueName: \"kubernetes.io/projected/08df06a8-008b-408b-b2cb-209f5cd2a5e5-kube-api-access-2vmhm\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.071504 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-utilities\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.071560 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-catalog-content\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.072005 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-utilities\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.072051 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-catalog-content\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.093307 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vmhm\" (UniqueName: \"kubernetes.io/projected/08df06a8-008b-408b-b2cb-209f5cd2a5e5-kube-api-access-2vmhm\") pod \"redhat-operators-f5jzk\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.211513 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.256989 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" event={"ID":"0430fbc5-3d73-4885-8670-61dd256714be","Type":"ContainerDied","Data":"f361c7f51064b41340e7a3f4e22e029585053c28a9d402e33e6ea496a817ff39"} Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.257040 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f361c7f51064b41340e7a3f4e22e029585053c28a9d402e33e6ea496a817ff39" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.257116 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.360509 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm"] Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.364744 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.369849 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.369953 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.369979 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.381771 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.395279 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.395656 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.395782 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.395936 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.397472 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm"] Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492671 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492747 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492773 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5xtd\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-kube-api-access-h5xtd\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492822 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492848 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492899 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492946 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.492982 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.493005 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.493043 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.493085 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.493122 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.493177 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.493226 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595573 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595643 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595668 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5xtd\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-kube-api-access-h5xtd\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595713 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595737 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595781 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595821 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595852 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595874 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595908 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595943 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.595973 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.596020 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.596064 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.630924 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.643434 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.644228 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.644364 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.645404 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.645568 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.645604 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.645630 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.647681 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.647704 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.653230 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.653791 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.654264 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.656992 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5xtd\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-kube-api-access-h5xtd\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.761826 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:52:30 crc kubenswrapper[4679]: I1202 10:52:30.954968 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f5jzk"] Dec 02 10:52:31 crc kubenswrapper[4679]: I1202 10:52:31.271458 4679 generic.go:334] "Generic (PLEG): container finished" podID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerID="5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8" exitCode=0 Dec 02 10:52:31 crc kubenswrapper[4679]: I1202 10:52:31.271513 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerDied","Data":"5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8"} Dec 02 10:52:31 crc kubenswrapper[4679]: I1202 10:52:31.271549 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerStarted","Data":"863e1f0613058059de767c9b0c33164faee40add91a465b0c9d954e4bb8ed6c9"} Dec 02 10:52:31 crc kubenswrapper[4679]: I1202 10:52:31.366488 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm"] Dec 02 10:52:31 crc kubenswrapper[4679]: W1202 10:52:31.369616 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca2fc1cb_e547_47c7_b27a_7de8429d2032.slice/crio-6aa15dd7f53fdcb98a80fe51c65fd8d5375de49eacb43a8db63e61e41c32302e WatchSource:0}: Error finding container 6aa15dd7f53fdcb98a80fe51c65fd8d5375de49eacb43a8db63e61e41c32302e: Status 404 returned error can't find the container with id 6aa15dd7f53fdcb98a80fe51c65fd8d5375de49eacb43a8db63e61e41c32302e Dec 02 10:52:32 crc kubenswrapper[4679]: I1202 10:52:32.285214 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" event={"ID":"ca2fc1cb-e547-47c7-b27a-7de8429d2032","Type":"ContainerStarted","Data":"08c38a38eecbdde6b26b27c2aae32f6800cc60a5fe3b9ae4b7f14b45705f1028"} Dec 02 10:52:32 crc kubenswrapper[4679]: I1202 10:52:32.285792 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" event={"ID":"ca2fc1cb-e547-47c7-b27a-7de8429d2032","Type":"ContainerStarted","Data":"6aa15dd7f53fdcb98a80fe51c65fd8d5375de49eacb43a8db63e61e41c32302e"} Dec 02 10:52:32 crc kubenswrapper[4679]: I1202 10:52:32.291093 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerStarted","Data":"4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40"} Dec 02 10:52:32 crc kubenswrapper[4679]: I1202 10:52:32.305710 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" podStartSLOduration=1.80568079 podStartE2EDuration="2.305692634s" podCreationTimestamp="2025-12-02 10:52:30 +0000 UTC" firstStartedPulling="2025-12-02 10:52:31.372821883 +0000 UTC m=+2004.702960743" lastFinishedPulling="2025-12-02 10:52:31.872833727 +0000 UTC m=+2005.202972587" observedRunningTime="2025-12-02 10:52:32.304533341 +0000 UTC m=+2005.634672221" watchObservedRunningTime="2025-12-02 10:52:32.305692634 +0000 UTC m=+2005.635831494" Dec 02 10:52:35 crc kubenswrapper[4679]: I1202 10:52:35.319581 4679 generic.go:334] "Generic (PLEG): container finished" podID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerID="4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40" exitCode=0 Dec 02 10:52:35 crc kubenswrapper[4679]: I1202 10:52:35.319653 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerDied","Data":"4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40"} Dec 02 10:52:35 crc kubenswrapper[4679]: E1202 10:52:35.330464 4679 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08df06a8_008b_408b_b2cb_209f5cd2a5e5.slice/crio-conmon-4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40.scope\": RecentStats: unable to find data in memory cache]" Dec 02 10:52:37 crc kubenswrapper[4679]: I1202 10:52:37.342894 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerStarted","Data":"1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699"} Dec 02 10:52:37 crc kubenswrapper[4679]: I1202 10:52:37.367500 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f5jzk" podStartSLOduration=3.303919434 podStartE2EDuration="8.367484723s" podCreationTimestamp="2025-12-02 10:52:29 +0000 UTC" firstStartedPulling="2025-12-02 10:52:31.27468434 +0000 UTC m=+2004.604823200" lastFinishedPulling="2025-12-02 10:52:36.338249629 +0000 UTC m=+2009.668388489" observedRunningTime="2025-12-02 10:52:37.365495686 +0000 UTC m=+2010.695634546" watchObservedRunningTime="2025-12-02 10:52:37.367484723 +0000 UTC m=+2010.697623583" Dec 02 10:52:40 crc kubenswrapper[4679]: I1202 10:52:40.211818 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:40 crc kubenswrapper[4679]: I1202 10:52:40.212289 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:41 crc kubenswrapper[4679]: I1202 10:52:41.254663 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f5jzk" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="registry-server" probeResult="failure" output=< Dec 02 10:52:41 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 10:52:41 crc kubenswrapper[4679]: > Dec 02 10:52:50 crc kubenswrapper[4679]: I1202 10:52:50.258264 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:50 crc kubenswrapper[4679]: I1202 10:52:50.309102 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:50 crc kubenswrapper[4679]: I1202 10:52:50.496695 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5jzk"] Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.458284 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f5jzk" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="registry-server" containerID="cri-o://1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699" gracePeriod=2 Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.910368 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.947611 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vmhm\" (UniqueName: \"kubernetes.io/projected/08df06a8-008b-408b-b2cb-209f5cd2a5e5-kube-api-access-2vmhm\") pod \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.947679 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-utilities\") pod \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.947705 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-catalog-content\") pod \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\" (UID: \"08df06a8-008b-408b-b2cb-209f5cd2a5e5\") " Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.948838 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-utilities" (OuterVolumeSpecName: "utilities") pod "08df06a8-008b-408b-b2cb-209f5cd2a5e5" (UID: "08df06a8-008b-408b-b2cb-209f5cd2a5e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:52:51 crc kubenswrapper[4679]: I1202 10:52:51.953812 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08df06a8-008b-408b-b2cb-209f5cd2a5e5-kube-api-access-2vmhm" (OuterVolumeSpecName: "kube-api-access-2vmhm") pod "08df06a8-008b-408b-b2cb-209f5cd2a5e5" (UID: "08df06a8-008b-408b-b2cb-209f5cd2a5e5"). InnerVolumeSpecName "kube-api-access-2vmhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.049776 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vmhm\" (UniqueName: \"kubernetes.io/projected/08df06a8-008b-408b-b2cb-209f5cd2a5e5-kube-api-access-2vmhm\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.049811 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.059473 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08df06a8-008b-408b-b2cb-209f5cd2a5e5" (UID: "08df06a8-008b-408b-b2cb-209f5cd2a5e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.151321 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08df06a8-008b-408b-b2cb-209f5cd2a5e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.469852 4679 generic.go:334] "Generic (PLEG): container finished" podID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerID="1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699" exitCode=0 Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.469899 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerDied","Data":"1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699"} Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.469938 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f5jzk" event={"ID":"08df06a8-008b-408b-b2cb-209f5cd2a5e5","Type":"ContainerDied","Data":"863e1f0613058059de767c9b0c33164faee40add91a465b0c9d954e4bb8ed6c9"} Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.469952 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f5jzk" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.469958 4679 scope.go:117] "RemoveContainer" containerID="1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.525172 4679 scope.go:117] "RemoveContainer" containerID="4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.551367 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f5jzk"] Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.558411 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f5jzk"] Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.622105 4679 scope.go:117] "RemoveContainer" containerID="5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.663039 4679 scope.go:117] "RemoveContainer" containerID="1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699" Dec 02 10:52:52 crc kubenswrapper[4679]: E1202 10:52:52.663686 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699\": container with ID starting with 1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699 not found: ID does not exist" containerID="1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.663713 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699"} err="failed to get container status \"1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699\": rpc error: code = NotFound desc = could not find container \"1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699\": container with ID starting with 1d7bab006cfc6ea2afe3b00fd1b2cb73c14b93a727a716b1af275e6410b89699 not found: ID does not exist" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.663735 4679 scope.go:117] "RemoveContainer" containerID="4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40" Dec 02 10:52:52 crc kubenswrapper[4679]: E1202 10:52:52.664204 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40\": container with ID starting with 4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40 not found: ID does not exist" containerID="4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.664220 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40"} err="failed to get container status \"4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40\": rpc error: code = NotFound desc = could not find container \"4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40\": container with ID starting with 4d1057a0fff254921d0b1e33470104394c1950061f89885c5790437fc6ff1c40 not found: ID does not exist" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.664232 4679 scope.go:117] "RemoveContainer" containerID="5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8" Dec 02 10:52:52 crc kubenswrapper[4679]: E1202 10:52:52.664476 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8\": container with ID starting with 5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8 not found: ID does not exist" containerID="5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.664511 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8"} err="failed to get container status \"5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8\": rpc error: code = NotFound desc = could not find container \"5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8\": container with ID starting with 5037bea28038e4d937fc2b0c7ff90df38405a4fc9c1e1bd0936d8fd4f79333a8 not found: ID does not exist" Dec 02 10:52:52 crc kubenswrapper[4679]: I1202 10:52:52.919998 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" path="/var/lib/kubelet/pods/08df06a8-008b-408b-b2cb-209f5cd2a5e5/volumes" Dec 02 10:53:10 crc kubenswrapper[4679]: I1202 10:53:10.625234 4679 generic.go:334] "Generic (PLEG): container finished" podID="ca2fc1cb-e547-47c7-b27a-7de8429d2032" containerID="08c38a38eecbdde6b26b27c2aae32f6800cc60a5fe3b9ae4b7f14b45705f1028" exitCode=0 Dec 02 10:53:10 crc kubenswrapper[4679]: I1202 10:53:10.625359 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" event={"ID":"ca2fc1cb-e547-47c7-b27a-7de8429d2032","Type":"ContainerDied","Data":"08c38a38eecbdde6b26b27c2aae32f6800cc60a5fe3b9ae4b7f14b45705f1028"} Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.038814 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132336 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-libvirt-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132409 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ssh-key\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132462 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ovn-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132518 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132580 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132629 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-inventory\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132688 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-neutron-metadata-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132736 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5xtd\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-kube-api-access-h5xtd\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132755 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132849 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-repo-setup-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132892 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-nova-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132919 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-bootstrap-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132938 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-ovn-default-certs-0\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.132987 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-telemetry-combined-ca-bundle\") pod \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\" (UID: \"ca2fc1cb-e547-47c7-b27a-7de8429d2032\") " Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.142165 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.142277 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-kube-api-access-h5xtd" (OuterVolumeSpecName: "kube-api-access-h5xtd") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "kube-api-access-h5xtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.142750 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.142832 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.142846 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.142971 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.143932 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.143969 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.144004 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.145979 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.149278 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.151178 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.167242 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-inventory" (OuterVolumeSpecName: "inventory") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.174102 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca2fc1cb-e547-47c7-b27a-7de8429d2032" (UID: "ca2fc1cb-e547-47c7-b27a-7de8429d2032"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234829 4679 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234868 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234882 4679 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234898 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234913 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234929 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234946 4679 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234960 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234973 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5xtd\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-kube-api-access-h5xtd\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.234986 4679 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.235001 4679 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.235014 4679 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.235027 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ca2fc1cb-e547-47c7-b27a-7de8429d2032-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.235040 4679 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2fc1cb-e547-47c7-b27a-7de8429d2032-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.645035 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" event={"ID":"ca2fc1cb-e547-47c7-b27a-7de8429d2032","Type":"ContainerDied","Data":"6aa15dd7f53fdcb98a80fe51c65fd8d5375de49eacb43a8db63e61e41c32302e"} Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.645082 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6aa15dd7f53fdcb98a80fe51c65fd8d5375de49eacb43a8db63e61e41c32302e" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.645159 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.746432 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr"] Dec 02 10:53:12 crc kubenswrapper[4679]: E1202 10:53:12.746810 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="extract-utilities" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.746829 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="extract-utilities" Dec 02 10:53:12 crc kubenswrapper[4679]: E1202 10:53:12.746852 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="extract-content" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.746859 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="extract-content" Dec 02 10:53:12 crc kubenswrapper[4679]: E1202 10:53:12.746870 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="registry-server" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.746876 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="registry-server" Dec 02 10:53:12 crc kubenswrapper[4679]: E1202 10:53:12.746900 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2fc1cb-e547-47c7-b27a-7de8429d2032" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.746908 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2fc1cb-e547-47c7-b27a-7de8429d2032" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.747089 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2fc1cb-e547-47c7-b27a-7de8429d2032" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.747103 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="08df06a8-008b-408b-b2cb-209f5cd2a5e5" containerName="registry-server" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.747774 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.749544 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.749802 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.749971 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.750369 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.765573 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.776445 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr"] Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.948740 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.948790 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.949016 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.949056 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knl68\" (UniqueName: \"kubernetes.io/projected/7e6fa9dc-0546-45fd-847c-29266ce64dfa-kube-api-access-knl68\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:12 crc kubenswrapper[4679]: I1202 10:53:12.949100 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.051234 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.051329 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knl68\" (UniqueName: \"kubernetes.io/projected/7e6fa9dc-0546-45fd-847c-29266ce64dfa-kube-api-access-knl68\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.051389 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.051536 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.051561 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.052755 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.055574 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.055840 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.056469 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.070513 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knl68\" (UniqueName: \"kubernetes.io/projected/7e6fa9dc-0546-45fd-847c-29266ce64dfa-kube-api-access-knl68\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gzdgr\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.077086 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.593512 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr"] Dec 02 10:53:13 crc kubenswrapper[4679]: I1202 10:53:13.654769 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" event={"ID":"7e6fa9dc-0546-45fd-847c-29266ce64dfa","Type":"ContainerStarted","Data":"cd7271a2c67c1dd3d6e086e851587b5b881ee998756a732431859e249726ebc2"} Dec 02 10:53:14 crc kubenswrapper[4679]: I1202 10:53:14.667525 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" event={"ID":"7e6fa9dc-0546-45fd-847c-29266ce64dfa","Type":"ContainerStarted","Data":"7e9959b81ec6a80ceb0fecfc4a81ca1a787fec10cbc60a7cea77a04a84891799"} Dec 02 10:53:14 crc kubenswrapper[4679]: I1202 10:53:14.695082 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" podStartSLOduration=2.2176002 podStartE2EDuration="2.695064977s" podCreationTimestamp="2025-12-02 10:53:12 +0000 UTC" firstStartedPulling="2025-12-02 10:53:13.60921377 +0000 UTC m=+2046.939352630" lastFinishedPulling="2025-12-02 10:53:14.086678547 +0000 UTC m=+2047.416817407" observedRunningTime="2025-12-02 10:53:14.687024637 +0000 UTC m=+2048.017163507" watchObservedRunningTime="2025-12-02 10:53:14.695064977 +0000 UTC m=+2048.025203837" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.741055 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wdrq2"] Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.743758 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.753248 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdrq2"] Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.892554 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-catalog-content\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.892588 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sg22\" (UniqueName: \"kubernetes.io/projected/34c2a702-a782-4da0-a9d6-4869d1b347b0-kube-api-access-8sg22\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.892625 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-utilities\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.994171 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-catalog-content\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.994209 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sg22\" (UniqueName: \"kubernetes.io/projected/34c2a702-a782-4da0-a9d6-4869d1b347b0-kube-api-access-8sg22\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.994254 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-utilities\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.995284 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-catalog-content\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:55 crc kubenswrapper[4679]: I1202 10:53:55.995357 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-utilities\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:56 crc kubenswrapper[4679]: I1202 10:53:56.016783 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sg22\" (UniqueName: \"kubernetes.io/projected/34c2a702-a782-4da0-a9d6-4869d1b347b0-kube-api-access-8sg22\") pod \"redhat-marketplace-wdrq2\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:56 crc kubenswrapper[4679]: I1202 10:53:56.070288 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:53:56 crc kubenswrapper[4679]: I1202 10:53:56.508678 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdrq2"] Dec 02 10:53:57 crc kubenswrapper[4679]: I1202 10:53:57.061584 4679 generic.go:334] "Generic (PLEG): container finished" podID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerID="21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087" exitCode=0 Dec 02 10:53:57 crc kubenswrapper[4679]: I1202 10:53:57.061662 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdrq2" event={"ID":"34c2a702-a782-4da0-a9d6-4869d1b347b0","Type":"ContainerDied","Data":"21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087"} Dec 02 10:53:57 crc kubenswrapper[4679]: I1202 10:53:57.061703 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdrq2" event={"ID":"34c2a702-a782-4da0-a9d6-4869d1b347b0","Type":"ContainerStarted","Data":"16434a5b4273ae2728636718a29c0f5792692704c225e8a7c6980d376a8284db"} Dec 02 10:53:59 crc kubenswrapper[4679]: I1202 10:53:59.088995 4679 generic.go:334] "Generic (PLEG): container finished" podID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerID="4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1" exitCode=0 Dec 02 10:53:59 crc kubenswrapper[4679]: I1202 10:53:59.089053 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdrq2" event={"ID":"34c2a702-a782-4da0-a9d6-4869d1b347b0","Type":"ContainerDied","Data":"4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1"} Dec 02 10:54:00 crc kubenswrapper[4679]: I1202 10:54:00.100992 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdrq2" event={"ID":"34c2a702-a782-4da0-a9d6-4869d1b347b0","Type":"ContainerStarted","Data":"7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3"} Dec 02 10:54:00 crc kubenswrapper[4679]: I1202 10:54:00.123177 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wdrq2" podStartSLOduration=2.543177669 podStartE2EDuration="5.123154676s" podCreationTimestamp="2025-12-02 10:53:55 +0000 UTC" firstStartedPulling="2025-12-02 10:53:57.064032704 +0000 UTC m=+2090.394171564" lastFinishedPulling="2025-12-02 10:53:59.644009711 +0000 UTC m=+2092.974148571" observedRunningTime="2025-12-02 10:54:00.12121495 +0000 UTC m=+2093.451353840" watchObservedRunningTime="2025-12-02 10:54:00.123154676 +0000 UTC m=+2093.453293546" Dec 02 10:54:06 crc kubenswrapper[4679]: I1202 10:54:06.071404 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:54:06 crc kubenswrapper[4679]: I1202 10:54:06.072044 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:54:06 crc kubenswrapper[4679]: I1202 10:54:06.141255 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:54:06 crc kubenswrapper[4679]: I1202 10:54:06.205828 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:54:06 crc kubenswrapper[4679]: I1202 10:54:06.378554 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdrq2"] Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.174539 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wdrq2" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="registry-server" containerID="cri-o://7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3" gracePeriod=2 Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.629926 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.753109 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-catalog-content\") pod \"34c2a702-a782-4da0-a9d6-4869d1b347b0\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.753379 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sg22\" (UniqueName: \"kubernetes.io/projected/34c2a702-a782-4da0-a9d6-4869d1b347b0-kube-api-access-8sg22\") pod \"34c2a702-a782-4da0-a9d6-4869d1b347b0\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.753479 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-utilities\") pod \"34c2a702-a782-4da0-a9d6-4869d1b347b0\" (UID: \"34c2a702-a782-4da0-a9d6-4869d1b347b0\") " Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.754260 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-utilities" (OuterVolumeSpecName: "utilities") pod "34c2a702-a782-4da0-a9d6-4869d1b347b0" (UID: "34c2a702-a782-4da0-a9d6-4869d1b347b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.759276 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c2a702-a782-4da0-a9d6-4869d1b347b0-kube-api-access-8sg22" (OuterVolumeSpecName: "kube-api-access-8sg22") pod "34c2a702-a782-4da0-a9d6-4869d1b347b0" (UID: "34c2a702-a782-4da0-a9d6-4869d1b347b0"). InnerVolumeSpecName "kube-api-access-8sg22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.771621 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34c2a702-a782-4da0-a9d6-4869d1b347b0" (UID: "34c2a702-a782-4da0-a9d6-4869d1b347b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.855801 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.855847 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sg22\" (UniqueName: \"kubernetes.io/projected/34c2a702-a782-4da0-a9d6-4869d1b347b0-kube-api-access-8sg22\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:08 crc kubenswrapper[4679]: I1202 10:54:08.855863 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2a702-a782-4da0-a9d6-4869d1b347b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.185689 4679 generic.go:334] "Generic (PLEG): container finished" podID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerID="7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3" exitCode=0 Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.185754 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdrq2" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.185770 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdrq2" event={"ID":"34c2a702-a782-4da0-a9d6-4869d1b347b0","Type":"ContainerDied","Data":"7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3"} Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.185837 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdrq2" event={"ID":"34c2a702-a782-4da0-a9d6-4869d1b347b0","Type":"ContainerDied","Data":"16434a5b4273ae2728636718a29c0f5792692704c225e8a7c6980d376a8284db"} Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.185874 4679 scope.go:117] "RemoveContainer" containerID="7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.209362 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdrq2"] Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.214268 4679 scope.go:117] "RemoveContainer" containerID="4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.216884 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdrq2"] Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.231412 4679 scope.go:117] "RemoveContainer" containerID="21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.279801 4679 scope.go:117] "RemoveContainer" containerID="7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3" Dec 02 10:54:09 crc kubenswrapper[4679]: E1202 10:54:09.285082 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3\": container with ID starting with 7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3 not found: ID does not exist" containerID="7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.285319 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3"} err="failed to get container status \"7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3\": rpc error: code = NotFound desc = could not find container \"7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3\": container with ID starting with 7e1a2d5ba5ead2008ac6466a72ab0e582fa87cd87121b9fd39a53f791c5ef1f3 not found: ID does not exist" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.285346 4679 scope.go:117] "RemoveContainer" containerID="4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1" Dec 02 10:54:09 crc kubenswrapper[4679]: E1202 10:54:09.285694 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1\": container with ID starting with 4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1 not found: ID does not exist" containerID="4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.285731 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1"} err="failed to get container status \"4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1\": rpc error: code = NotFound desc = could not find container \"4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1\": container with ID starting with 4cd2df9b0e91a14421fb37a9688a2e258a0a205fc2dea80c91a49cb7a5d668f1 not found: ID does not exist" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.285756 4679 scope.go:117] "RemoveContainer" containerID="21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087" Dec 02 10:54:09 crc kubenswrapper[4679]: E1202 10:54:09.286036 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087\": container with ID starting with 21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087 not found: ID does not exist" containerID="21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087" Dec 02 10:54:09 crc kubenswrapper[4679]: I1202 10:54:09.286059 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087"} err="failed to get container status \"21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087\": rpc error: code = NotFound desc = could not find container \"21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087\": container with ID starting with 21049d1772354230bf97ada7628d001d7e20dd9a4889977cefed0cd582bf0087 not found: ID does not exist" Dec 02 10:54:10 crc kubenswrapper[4679]: I1202 10:54:10.924096 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" path="/var/lib/kubelet/pods/34c2a702-a782-4da0-a9d6-4869d1b347b0/volumes" Dec 02 10:54:17 crc kubenswrapper[4679]: I1202 10:54:17.272188 4679 generic.go:334] "Generic (PLEG): container finished" podID="7e6fa9dc-0546-45fd-847c-29266ce64dfa" containerID="7e9959b81ec6a80ceb0fecfc4a81ca1a787fec10cbc60a7cea77a04a84891799" exitCode=0 Dec 02 10:54:17 crc kubenswrapper[4679]: I1202 10:54:17.272268 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" event={"ID":"7e6fa9dc-0546-45fd-847c-29266ce64dfa","Type":"ContainerDied","Data":"7e9959b81ec6a80ceb0fecfc4a81ca1a787fec10cbc60a7cea77a04a84891799"} Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.777396 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.867475 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovncontroller-config-0\") pod \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.867530 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knl68\" (UniqueName: \"kubernetes.io/projected/7e6fa9dc-0546-45fd-847c-29266ce64dfa-kube-api-access-knl68\") pod \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.867563 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ssh-key\") pod \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.867648 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovn-combined-ca-bundle\") pod \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.868669 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-inventory\") pod \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\" (UID: \"7e6fa9dc-0546-45fd-847c-29266ce64dfa\") " Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.873911 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e6fa9dc-0546-45fd-847c-29266ce64dfa-kube-api-access-knl68" (OuterVolumeSpecName: "kube-api-access-knl68") pod "7e6fa9dc-0546-45fd-847c-29266ce64dfa" (UID: "7e6fa9dc-0546-45fd-847c-29266ce64dfa"). InnerVolumeSpecName "kube-api-access-knl68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.874476 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7e6fa9dc-0546-45fd-847c-29266ce64dfa" (UID: "7e6fa9dc-0546-45fd-847c-29266ce64dfa"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.893754 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "7e6fa9dc-0546-45fd-847c-29266ce64dfa" (UID: "7e6fa9dc-0546-45fd-847c-29266ce64dfa"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.900090 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-inventory" (OuterVolumeSpecName: "inventory") pod "7e6fa9dc-0546-45fd-847c-29266ce64dfa" (UID: "7e6fa9dc-0546-45fd-847c-29266ce64dfa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.900637 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7e6fa9dc-0546-45fd-847c-29266ce64dfa" (UID: "7e6fa9dc-0546-45fd-847c-29266ce64dfa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.970802 4679 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.970909 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knl68\" (UniqueName: \"kubernetes.io/projected/7e6fa9dc-0546-45fd-847c-29266ce64dfa-kube-api-access-knl68\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.970920 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.970929 4679 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:18 crc kubenswrapper[4679]: I1202 10:54:18.970941 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e6fa9dc-0546-45fd-847c-29266ce64dfa-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.291885 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" event={"ID":"7e6fa9dc-0546-45fd-847c-29266ce64dfa","Type":"ContainerDied","Data":"cd7271a2c67c1dd3d6e086e851587b5b881ee998756a732431859e249726ebc2"} Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.292281 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd7271a2c67c1dd3d6e086e851587b5b881ee998756a732431859e249726ebc2" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.291960 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gzdgr" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.404996 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r"] Dec 02 10:54:19 crc kubenswrapper[4679]: E1202 10:54:19.405376 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="extract-content" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.405394 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="extract-content" Dec 02 10:54:19 crc kubenswrapper[4679]: E1202 10:54:19.405437 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e6fa9dc-0546-45fd-847c-29266ce64dfa" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.405444 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e6fa9dc-0546-45fd-847c-29266ce64dfa" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 10:54:19 crc kubenswrapper[4679]: E1202 10:54:19.405456 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="registry-server" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.405462 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="registry-server" Dec 02 10:54:19 crc kubenswrapper[4679]: E1202 10:54:19.405472 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="extract-utilities" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.405478 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="extract-utilities" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.405645 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c2a702-a782-4da0-a9d6-4869d1b347b0" containerName="registry-server" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.405667 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e6fa9dc-0546-45fd-847c-29266ce64dfa" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.406253 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.408964 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.409431 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.409459 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.409506 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.409433 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.409543 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.415199 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r"] Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.480004 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.480046 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.480065 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.480102 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.480271 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8hlq\" (UniqueName: \"kubernetes.io/projected/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-kube-api-access-p8hlq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.480323 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.582147 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.582221 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.582251 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.582328 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.582443 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8hlq\" (UniqueName: \"kubernetes.io/projected/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-kube-api-access-p8hlq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.582468 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.587444 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.587873 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.588844 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.590008 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.594075 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.601665 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8hlq\" (UniqueName: \"kubernetes.io/projected/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-kube-api-access-p8hlq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:19 crc kubenswrapper[4679]: I1202 10:54:19.732345 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:54:20 crc kubenswrapper[4679]: I1202 10:54:20.276627 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r"] Dec 02 10:54:20 crc kubenswrapper[4679]: W1202 10:54:20.278811 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7764127f_88d3_4c9e_9eb2_9a6c37585ca3.slice/crio-0333fc3d73ebddc7f47cb6c8d23931bba37964e29c4936ddbc0013b5f33f70af WatchSource:0}: Error finding container 0333fc3d73ebddc7f47cb6c8d23931bba37964e29c4936ddbc0013b5f33f70af: Status 404 returned error can't find the container with id 0333fc3d73ebddc7f47cb6c8d23931bba37964e29c4936ddbc0013b5f33f70af Dec 02 10:54:20 crc kubenswrapper[4679]: I1202 10:54:20.302319 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" event={"ID":"7764127f-88d3-4c9e-9eb2-9a6c37585ca3","Type":"ContainerStarted","Data":"0333fc3d73ebddc7f47cb6c8d23931bba37964e29c4936ddbc0013b5f33f70af"} Dec 02 10:54:21 crc kubenswrapper[4679]: I1202 10:54:21.312373 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" event={"ID":"7764127f-88d3-4c9e-9eb2-9a6c37585ca3","Type":"ContainerStarted","Data":"0f5f275fa837c5f1a960d52f60fa37a4ec43db52d8c548121cb45ea212fdca4b"} Dec 02 10:54:21 crc kubenswrapper[4679]: I1202 10:54:21.330440 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" podStartSLOduration=1.737354992 podStartE2EDuration="2.330425013s" podCreationTimestamp="2025-12-02 10:54:19 +0000 UTC" firstStartedPulling="2025-12-02 10:54:20.281743772 +0000 UTC m=+2113.611882632" lastFinishedPulling="2025-12-02 10:54:20.874813793 +0000 UTC m=+2114.204952653" observedRunningTime="2025-12-02 10:54:21.327717775 +0000 UTC m=+2114.657856635" watchObservedRunningTime="2025-12-02 10:54:21.330425013 +0000 UTC m=+2114.660563873" Dec 02 10:54:46 crc kubenswrapper[4679]: I1202 10:54:46.932172 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:54:46 crc kubenswrapper[4679]: I1202 10:54:46.932725 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:55:07 crc kubenswrapper[4679]: I1202 10:55:07.750367 4679 generic.go:334] "Generic (PLEG): container finished" podID="7764127f-88d3-4c9e-9eb2-9a6c37585ca3" containerID="0f5f275fa837c5f1a960d52f60fa37a4ec43db52d8c548121cb45ea212fdca4b" exitCode=0 Dec 02 10:55:07 crc kubenswrapper[4679]: I1202 10:55:07.750552 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" event={"ID":"7764127f-88d3-4c9e-9eb2-9a6c37585ca3","Type":"ContainerDied","Data":"0f5f275fa837c5f1a960d52f60fa37a4ec43db52d8c548121cb45ea212fdca4b"} Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.234656 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.346959 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.347007 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-ssh-key\") pod \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.347040 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8hlq\" (UniqueName: \"kubernetes.io/projected/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-kube-api-access-p8hlq\") pod \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.347145 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-inventory\") pod \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.347224 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-nova-metadata-neutron-config-0\") pod \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.347372 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-metadata-combined-ca-bundle\") pod \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\" (UID: \"7764127f-88d3-4c9e-9eb2-9a6c37585ca3\") " Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.354805 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-kube-api-access-p8hlq" (OuterVolumeSpecName: "kube-api-access-p8hlq") pod "7764127f-88d3-4c9e-9eb2-9a6c37585ca3" (UID: "7764127f-88d3-4c9e-9eb2-9a6c37585ca3"). InnerVolumeSpecName "kube-api-access-p8hlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.356110 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7764127f-88d3-4c9e-9eb2-9a6c37585ca3" (UID: "7764127f-88d3-4c9e-9eb2-9a6c37585ca3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.381231 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "7764127f-88d3-4c9e-9eb2-9a6c37585ca3" (UID: "7764127f-88d3-4c9e-9eb2-9a6c37585ca3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.383983 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "7764127f-88d3-4c9e-9eb2-9a6c37585ca3" (UID: "7764127f-88d3-4c9e-9eb2-9a6c37585ca3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.384739 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7764127f-88d3-4c9e-9eb2-9a6c37585ca3" (UID: "7764127f-88d3-4c9e-9eb2-9a6c37585ca3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.398541 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-inventory" (OuterVolumeSpecName: "inventory") pod "7764127f-88d3-4c9e-9eb2-9a6c37585ca3" (UID: "7764127f-88d3-4c9e-9eb2-9a6c37585ca3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.450336 4679 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.450374 4679 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.450391 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.450405 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8hlq\" (UniqueName: \"kubernetes.io/projected/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-kube-api-access-p8hlq\") on node \"crc\" DevicePath \"\"" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.450418 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.450429 4679 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7764127f-88d3-4c9e-9eb2-9a6c37585ca3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.769719 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" event={"ID":"7764127f-88d3-4c9e-9eb2-9a6c37585ca3","Type":"ContainerDied","Data":"0333fc3d73ebddc7f47cb6c8d23931bba37964e29c4936ddbc0013b5f33f70af"} Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.769760 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0333fc3d73ebddc7f47cb6c8d23931bba37964e29c4936ddbc0013b5f33f70af" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.769807 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.879595 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76"] Dec 02 10:55:09 crc kubenswrapper[4679]: E1202 10:55:09.880592 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7764127f-88d3-4c9e-9eb2-9a6c37585ca3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.880616 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="7764127f-88d3-4c9e-9eb2-9a6c37585ca3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.880896 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="7764127f-88d3-4c9e-9eb2-9a6c37585ca3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.881671 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.886828 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.888483 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.888678 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.888840 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.889504 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.916101 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76"] Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.958461 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghscj\" (UniqueName: \"kubernetes.io/projected/1202831e-ebbe-4507-b89a-b3e9de9d64e6-kube-api-access-ghscj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.958737 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.958927 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.958967 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:09 crc kubenswrapper[4679]: I1202 10:55:09.959029 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.060727 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghscj\" (UniqueName: \"kubernetes.io/projected/1202831e-ebbe-4507-b89a-b3e9de9d64e6-kube-api-access-ghscj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.061088 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.061396 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.062138 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.062530 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.066237 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.066721 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.067717 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.069495 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.082634 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghscj\" (UniqueName: \"kubernetes.io/projected/1202831e-ebbe-4507-b89a-b3e9de9d64e6-kube-api-access-ghscj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lhf76\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.221832 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.765325 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76"] Dec 02 10:55:10 crc kubenswrapper[4679]: I1202 10:55:10.778811 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" event={"ID":"1202831e-ebbe-4507-b89a-b3e9de9d64e6","Type":"ContainerStarted","Data":"eff0d9abae8907883ebb8082dd825ab85483c894e4c88701cc2acc7d40adec12"} Dec 02 10:55:12 crc kubenswrapper[4679]: I1202 10:55:12.820995 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" event={"ID":"1202831e-ebbe-4507-b89a-b3e9de9d64e6","Type":"ContainerStarted","Data":"59edbc2c679fce915a02de73345a1453c1bfded9cb1fff4df1c3b76def6c157f"} Dec 02 10:55:12 crc kubenswrapper[4679]: I1202 10:55:12.857833 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" podStartSLOduration=3.045360956 podStartE2EDuration="3.857810276s" podCreationTimestamp="2025-12-02 10:55:09 +0000 UTC" firstStartedPulling="2025-12-02 10:55:10.768884645 +0000 UTC m=+2164.099023505" lastFinishedPulling="2025-12-02 10:55:11.581333965 +0000 UTC m=+2164.911472825" observedRunningTime="2025-12-02 10:55:12.840656574 +0000 UTC m=+2166.170795434" watchObservedRunningTime="2025-12-02 10:55:12.857810276 +0000 UTC m=+2166.187949136" Dec 02 10:55:16 crc kubenswrapper[4679]: I1202 10:55:16.932327 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:55:16 crc kubenswrapper[4679]: I1202 10:55:16.933607 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:55:46 crc kubenswrapper[4679]: I1202 10:55:46.931827 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:55:46 crc kubenswrapper[4679]: I1202 10:55:46.932492 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:55:46 crc kubenswrapper[4679]: I1202 10:55:46.932545 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:55:46 crc kubenswrapper[4679]: I1202 10:55:46.933192 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c8b1bf853618b96cd693f0b2fd8cdfb8559d98bac63a4c138f60ed0d408a7d0"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:55:46 crc kubenswrapper[4679]: I1202 10:55:46.933254 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://9c8b1bf853618b96cd693f0b2fd8cdfb8559d98bac63a4c138f60ed0d408a7d0" gracePeriod=600 Dec 02 10:55:47 crc kubenswrapper[4679]: I1202 10:55:47.451736 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="9c8b1bf853618b96cd693f0b2fd8cdfb8559d98bac63a4c138f60ed0d408a7d0" exitCode=0 Dec 02 10:55:47 crc kubenswrapper[4679]: I1202 10:55:47.451812 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"9c8b1bf853618b96cd693f0b2fd8cdfb8559d98bac63a4c138f60ed0d408a7d0"} Dec 02 10:55:47 crc kubenswrapper[4679]: I1202 10:55:47.452130 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65"} Dec 02 10:55:47 crc kubenswrapper[4679]: I1202 10:55:47.452161 4679 scope.go:117] "RemoveContainer" containerID="df2d4f915c3f3ff743acb57289997745d572b11ec2756552f6fb17695e9ebc1a" Dec 02 10:57:28 crc kubenswrapper[4679]: I1202 10:57:28.932898 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwtzq"] Dec 02 10:57:28 crc kubenswrapper[4679]: I1202 10:57:28.936906 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwtzq"] Dec 02 10:57:28 crc kubenswrapper[4679]: I1202 10:57:28.937054 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:28 crc kubenswrapper[4679]: I1202 10:57:28.960296 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-catalog-content\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:28 crc kubenswrapper[4679]: I1202 10:57:28.960463 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-utilities\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:28 crc kubenswrapper[4679]: I1202 10:57:28.960500 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmj7r\" (UniqueName: \"kubernetes.io/projected/e0446c3f-fde3-4868-9bf3-6421d5e651ea-kube-api-access-gmj7r\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.062780 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-utilities\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.063448 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmj7r\" (UniqueName: \"kubernetes.io/projected/e0446c3f-fde3-4868-9bf3-6421d5e651ea-kube-api-access-gmj7r\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.063621 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-catalog-content\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.063366 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-utilities\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.064092 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-catalog-content\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.087331 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmj7r\" (UniqueName: \"kubernetes.io/projected/e0446c3f-fde3-4868-9bf3-6421d5e651ea-kube-api-access-gmj7r\") pod \"community-operators-zwtzq\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.269531 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:29 crc kubenswrapper[4679]: I1202 10:57:29.804287 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwtzq"] Dec 02 10:57:30 crc kubenswrapper[4679]: I1202 10:57:30.795030 4679 generic.go:334] "Generic (PLEG): container finished" podID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerID="44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d" exitCode=0 Dec 02 10:57:30 crc kubenswrapper[4679]: I1202 10:57:30.795086 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerDied","Data":"44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d"} Dec 02 10:57:30 crc kubenswrapper[4679]: I1202 10:57:30.795403 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerStarted","Data":"d001115f6ca4569e7ee3c527ee12b3f174ea7e7a7e56af8eabc896a063f9f1f3"} Dec 02 10:57:30 crc kubenswrapper[4679]: I1202 10:57:30.797647 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 10:57:31 crc kubenswrapper[4679]: I1202 10:57:31.807204 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerStarted","Data":"6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738"} Dec 02 10:57:32 crc kubenswrapper[4679]: I1202 10:57:32.820982 4679 generic.go:334] "Generic (PLEG): container finished" podID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerID="6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738" exitCode=0 Dec 02 10:57:32 crc kubenswrapper[4679]: I1202 10:57:32.821072 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerDied","Data":"6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738"} Dec 02 10:57:33 crc kubenswrapper[4679]: I1202 10:57:33.833021 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerStarted","Data":"7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9"} Dec 02 10:57:33 crc kubenswrapper[4679]: I1202 10:57:33.858201 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwtzq" podStartSLOduration=3.330898173 podStartE2EDuration="5.858178741s" podCreationTimestamp="2025-12-02 10:57:28 +0000 UTC" firstStartedPulling="2025-12-02 10:57:30.797299081 +0000 UTC m=+2304.127437981" lastFinishedPulling="2025-12-02 10:57:33.324579699 +0000 UTC m=+2306.654718549" observedRunningTime="2025-12-02 10:57:33.847782255 +0000 UTC m=+2307.177921125" watchObservedRunningTime="2025-12-02 10:57:33.858178741 +0000 UTC m=+2307.188317601" Dec 02 10:57:39 crc kubenswrapper[4679]: I1202 10:57:39.270165 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:39 crc kubenswrapper[4679]: I1202 10:57:39.270790 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:39 crc kubenswrapper[4679]: I1202 10:57:39.323904 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:39 crc kubenswrapper[4679]: I1202 10:57:39.933060 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:39 crc kubenswrapper[4679]: I1202 10:57:39.981777 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwtzq"] Dec 02 10:57:41 crc kubenswrapper[4679]: I1202 10:57:41.900237 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwtzq" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="registry-server" containerID="cri-o://7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9" gracePeriod=2 Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.367944 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.410661 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-catalog-content\") pod \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.410848 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-utilities\") pod \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.410895 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmj7r\" (UniqueName: \"kubernetes.io/projected/e0446c3f-fde3-4868-9bf3-6421d5e651ea-kube-api-access-gmj7r\") pod \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\" (UID: \"e0446c3f-fde3-4868-9bf3-6421d5e651ea\") " Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.413541 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-utilities" (OuterVolumeSpecName: "utilities") pod "e0446c3f-fde3-4868-9bf3-6421d5e651ea" (UID: "e0446c3f-fde3-4868-9bf3-6421d5e651ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.420579 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0446c3f-fde3-4868-9bf3-6421d5e651ea-kube-api-access-gmj7r" (OuterVolumeSpecName: "kube-api-access-gmj7r") pod "e0446c3f-fde3-4868-9bf3-6421d5e651ea" (UID: "e0446c3f-fde3-4868-9bf3-6421d5e651ea"). InnerVolumeSpecName "kube-api-access-gmj7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.473708 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0446c3f-fde3-4868-9bf3-6421d5e651ea" (UID: "e0446c3f-fde3-4868-9bf3-6421d5e651ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.513243 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.513278 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0446c3f-fde3-4868-9bf3-6421d5e651ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.513288 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmj7r\" (UniqueName: \"kubernetes.io/projected/e0446c3f-fde3-4868-9bf3-6421d5e651ea-kube-api-access-gmj7r\") on node \"crc\" DevicePath \"\"" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.909854 4679 generic.go:334] "Generic (PLEG): container finished" podID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerID="7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9" exitCode=0 Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.909955 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwtzq" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.920876 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerDied","Data":"7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9"} Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.920917 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwtzq" event={"ID":"e0446c3f-fde3-4868-9bf3-6421d5e651ea","Type":"ContainerDied","Data":"d001115f6ca4569e7ee3c527ee12b3f174ea7e7a7e56af8eabc896a063f9f1f3"} Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.920936 4679 scope.go:117] "RemoveContainer" containerID="7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.946097 4679 scope.go:117] "RemoveContainer" containerID="6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738" Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.949175 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwtzq"] Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.957537 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwtzq"] Dec 02 10:57:42 crc kubenswrapper[4679]: I1202 10:57:42.971671 4679 scope.go:117] "RemoveContainer" containerID="44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d" Dec 02 10:57:43 crc kubenswrapper[4679]: I1202 10:57:43.005169 4679 scope.go:117] "RemoveContainer" containerID="7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9" Dec 02 10:57:43 crc kubenswrapper[4679]: E1202 10:57:43.005718 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9\": container with ID starting with 7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9 not found: ID does not exist" containerID="7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9" Dec 02 10:57:43 crc kubenswrapper[4679]: I1202 10:57:43.005758 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9"} err="failed to get container status \"7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9\": rpc error: code = NotFound desc = could not find container \"7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9\": container with ID starting with 7500130a1892b8aa5a0ad7dcb3c9d8ef7c6caac835c8352ee98725799fd2d3e9 not found: ID does not exist" Dec 02 10:57:43 crc kubenswrapper[4679]: I1202 10:57:43.005785 4679 scope.go:117] "RemoveContainer" containerID="6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738" Dec 02 10:57:43 crc kubenswrapper[4679]: E1202 10:57:43.006180 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738\": container with ID starting with 6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738 not found: ID does not exist" containerID="6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738" Dec 02 10:57:43 crc kubenswrapper[4679]: I1202 10:57:43.006224 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738"} err="failed to get container status \"6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738\": rpc error: code = NotFound desc = could not find container \"6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738\": container with ID starting with 6088e49a50f4fdf77beb649daaaffdc8c94e0c754ad6b4be7935d826818b0738 not found: ID does not exist" Dec 02 10:57:43 crc kubenswrapper[4679]: I1202 10:57:43.006253 4679 scope.go:117] "RemoveContainer" containerID="44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d" Dec 02 10:57:43 crc kubenswrapper[4679]: E1202 10:57:43.006713 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d\": container with ID starting with 44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d not found: ID does not exist" containerID="44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d" Dec 02 10:57:43 crc kubenswrapper[4679]: I1202 10:57:43.006759 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d"} err="failed to get container status \"44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d\": rpc error: code = NotFound desc = could not find container \"44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d\": container with ID starting with 44a8fd4281c72156c3d6a200c1c6ae5c37ed694b6b74b2e711948bc417f7ad8d not found: ID does not exist" Dec 02 10:57:44 crc kubenswrapper[4679]: I1202 10:57:44.919066 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" path="/var/lib/kubelet/pods/e0446c3f-fde3-4868-9bf3-6421d5e651ea/volumes" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.900104 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hs9dh"] Dec 02 10:57:47 crc kubenswrapper[4679]: E1202 10:57:47.901011 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="registry-server" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.901030 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="registry-server" Dec 02 10:57:47 crc kubenswrapper[4679]: E1202 10:57:47.901040 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="extract-utilities" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.901047 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="extract-utilities" Dec 02 10:57:47 crc kubenswrapper[4679]: E1202 10:57:47.901078 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="extract-content" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.901089 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="extract-content" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.901393 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0446c3f-fde3-4868-9bf3-6421d5e651ea" containerName="registry-server" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.903238 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:47 crc kubenswrapper[4679]: I1202 10:57:47.914336 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hs9dh"] Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.012159 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-catalog-content\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.012627 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p82nz\" (UniqueName: \"kubernetes.io/projected/ee66f902-70c0-4953-b6ed-021772b1d0fc-kube-api-access-p82nz\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.012695 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-utilities\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.114277 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p82nz\" (UniqueName: \"kubernetes.io/projected/ee66f902-70c0-4953-b6ed-021772b1d0fc-kube-api-access-p82nz\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.114595 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-utilities\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.114715 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-catalog-content\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.115212 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-utilities\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.115263 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-catalog-content\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.148115 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p82nz\" (UniqueName: \"kubernetes.io/projected/ee66f902-70c0-4953-b6ed-021772b1d0fc-kube-api-access-p82nz\") pod \"certified-operators-hs9dh\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.228740 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.801289 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hs9dh"] Dec 02 10:57:48 crc kubenswrapper[4679]: I1202 10:57:48.974358 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hs9dh" event={"ID":"ee66f902-70c0-4953-b6ed-021772b1d0fc","Type":"ContainerStarted","Data":"4dcd2ac02423b88a4ba32b3ff05c4826ca00a3e5ca11f441aeafe83a2ab812f2"} Dec 02 10:57:49 crc kubenswrapper[4679]: I1202 10:57:49.982471 4679 generic.go:334] "Generic (PLEG): container finished" podID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerID="ffdb058ce0a18a62758504cffa6c80344f9bb431e157e4bd8ee20708e9fde971" exitCode=0 Dec 02 10:57:49 crc kubenswrapper[4679]: I1202 10:57:49.982516 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hs9dh" event={"ID":"ee66f902-70c0-4953-b6ed-021772b1d0fc","Type":"ContainerDied","Data":"ffdb058ce0a18a62758504cffa6c80344f9bb431e157e4bd8ee20708e9fde971"} Dec 02 10:57:52 crc kubenswrapper[4679]: I1202 10:57:52.001966 4679 generic.go:334] "Generic (PLEG): container finished" podID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerID="fbc4d14a801f59cbd15f0a35aa3c40310bd51730c36116adf767f14eee5cfc91" exitCode=0 Dec 02 10:57:52 crc kubenswrapper[4679]: I1202 10:57:52.002049 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hs9dh" event={"ID":"ee66f902-70c0-4953-b6ed-021772b1d0fc","Type":"ContainerDied","Data":"fbc4d14a801f59cbd15f0a35aa3c40310bd51730c36116adf767f14eee5cfc91"} Dec 02 10:57:53 crc kubenswrapper[4679]: I1202 10:57:53.015719 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hs9dh" event={"ID":"ee66f902-70c0-4953-b6ed-021772b1d0fc","Type":"ContainerStarted","Data":"b4ad0b43280ba20e037d81b48e420683ce84dfd2518ef5e732c6d0970f6151b6"} Dec 02 10:57:53 crc kubenswrapper[4679]: I1202 10:57:53.039473 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hs9dh" podStartSLOduration=3.559861809 podStartE2EDuration="6.039457477s" podCreationTimestamp="2025-12-02 10:57:47 +0000 UTC" firstStartedPulling="2025-12-02 10:57:49.98464965 +0000 UTC m=+2323.314788510" lastFinishedPulling="2025-12-02 10:57:52.464245318 +0000 UTC m=+2325.794384178" observedRunningTime="2025-12-02 10:57:53.037157341 +0000 UTC m=+2326.367296211" watchObservedRunningTime="2025-12-02 10:57:53.039457477 +0000 UTC m=+2326.369596337" Dec 02 10:57:58 crc kubenswrapper[4679]: I1202 10:57:58.229928 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:58 crc kubenswrapper[4679]: I1202 10:57:58.230864 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:58 crc kubenswrapper[4679]: I1202 10:57:58.281712 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:59 crc kubenswrapper[4679]: I1202 10:57:59.151104 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:57:59 crc kubenswrapper[4679]: I1202 10:57:59.523070 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hs9dh"] Dec 02 10:58:01 crc kubenswrapper[4679]: I1202 10:58:01.100118 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hs9dh" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="registry-server" containerID="cri-o://b4ad0b43280ba20e037d81b48e420683ce84dfd2518ef5e732c6d0970f6151b6" gracePeriod=2 Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.121281 4679 generic.go:334] "Generic (PLEG): container finished" podID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerID="b4ad0b43280ba20e037d81b48e420683ce84dfd2518ef5e732c6d0970f6151b6" exitCode=0 Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.121935 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hs9dh" event={"ID":"ee66f902-70c0-4953-b6ed-021772b1d0fc","Type":"ContainerDied","Data":"b4ad0b43280ba20e037d81b48e420683ce84dfd2518ef5e732c6d0970f6151b6"} Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.290364 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.398818 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-utilities\") pod \"ee66f902-70c0-4953-b6ed-021772b1d0fc\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.398952 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p82nz\" (UniqueName: \"kubernetes.io/projected/ee66f902-70c0-4953-b6ed-021772b1d0fc-kube-api-access-p82nz\") pod \"ee66f902-70c0-4953-b6ed-021772b1d0fc\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.399043 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-catalog-content\") pod \"ee66f902-70c0-4953-b6ed-021772b1d0fc\" (UID: \"ee66f902-70c0-4953-b6ed-021772b1d0fc\") " Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.402561 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-utilities" (OuterVolumeSpecName: "utilities") pod "ee66f902-70c0-4953-b6ed-021772b1d0fc" (UID: "ee66f902-70c0-4953-b6ed-021772b1d0fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.415636 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee66f902-70c0-4953-b6ed-021772b1d0fc-kube-api-access-p82nz" (OuterVolumeSpecName: "kube-api-access-p82nz") pod "ee66f902-70c0-4953-b6ed-021772b1d0fc" (UID: "ee66f902-70c0-4953-b6ed-021772b1d0fc"). InnerVolumeSpecName "kube-api-access-p82nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.457006 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee66f902-70c0-4953-b6ed-021772b1d0fc" (UID: "ee66f902-70c0-4953-b6ed-021772b1d0fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.501855 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.501902 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p82nz\" (UniqueName: \"kubernetes.io/projected/ee66f902-70c0-4953-b6ed-021772b1d0fc-kube-api-access-p82nz\") on node \"crc\" DevicePath \"\"" Dec 02 10:58:02 crc kubenswrapper[4679]: I1202 10:58:02.501917 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee66f902-70c0-4953-b6ed-021772b1d0fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.133818 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hs9dh" event={"ID":"ee66f902-70c0-4953-b6ed-021772b1d0fc","Type":"ContainerDied","Data":"4dcd2ac02423b88a4ba32b3ff05c4826ca00a3e5ca11f441aeafe83a2ab812f2"} Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.133873 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hs9dh" Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.134124 4679 scope.go:117] "RemoveContainer" containerID="b4ad0b43280ba20e037d81b48e420683ce84dfd2518ef5e732c6d0970f6151b6" Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.160018 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hs9dh"] Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.164126 4679 scope.go:117] "RemoveContainer" containerID="fbc4d14a801f59cbd15f0a35aa3c40310bd51730c36116adf767f14eee5cfc91" Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.169119 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hs9dh"] Dec 02 10:58:03 crc kubenswrapper[4679]: I1202 10:58:03.190530 4679 scope.go:117] "RemoveContainer" containerID="ffdb058ce0a18a62758504cffa6c80344f9bb431e157e4bd8ee20708e9fde971" Dec 02 10:58:04 crc kubenswrapper[4679]: I1202 10:58:04.921443 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" path="/var/lib/kubelet/pods/ee66f902-70c0-4953-b6ed-021772b1d0fc/volumes" Dec 02 10:58:16 crc kubenswrapper[4679]: I1202 10:58:16.931737 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:58:16 crc kubenswrapper[4679]: I1202 10:58:16.932701 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:58:46 crc kubenswrapper[4679]: I1202 10:58:46.931820 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:58:46 crc kubenswrapper[4679]: I1202 10:58:46.932280 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:59:14 crc kubenswrapper[4679]: I1202 10:59:14.835834 4679 generic.go:334] "Generic (PLEG): container finished" podID="1202831e-ebbe-4507-b89a-b3e9de9d64e6" containerID="59edbc2c679fce915a02de73345a1453c1bfded9cb1fff4df1c3b76def6c157f" exitCode=0 Dec 02 10:59:14 crc kubenswrapper[4679]: I1202 10:59:14.835910 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" event={"ID":"1202831e-ebbe-4507-b89a-b3e9de9d64e6","Type":"ContainerDied","Data":"59edbc2c679fce915a02de73345a1453c1bfded9cb1fff4df1c3b76def6c157f"} Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.275874 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.379973 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-secret-0\") pod \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.380057 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghscj\" (UniqueName: \"kubernetes.io/projected/1202831e-ebbe-4507-b89a-b3e9de9d64e6-kube-api-access-ghscj\") pod \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.380155 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-combined-ca-bundle\") pod \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.380212 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-ssh-key\") pod \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.380233 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-inventory\") pod \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\" (UID: \"1202831e-ebbe-4507-b89a-b3e9de9d64e6\") " Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.392856 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1202831e-ebbe-4507-b89a-b3e9de9d64e6-kube-api-access-ghscj" (OuterVolumeSpecName: "kube-api-access-ghscj") pod "1202831e-ebbe-4507-b89a-b3e9de9d64e6" (UID: "1202831e-ebbe-4507-b89a-b3e9de9d64e6"). InnerVolumeSpecName "kube-api-access-ghscj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.434448 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1202831e-ebbe-4507-b89a-b3e9de9d64e6" (UID: "1202831e-ebbe-4507-b89a-b3e9de9d64e6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.441067 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "1202831e-ebbe-4507-b89a-b3e9de9d64e6" (UID: "1202831e-ebbe-4507-b89a-b3e9de9d64e6"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.443033 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1202831e-ebbe-4507-b89a-b3e9de9d64e6" (UID: "1202831e-ebbe-4507-b89a-b3e9de9d64e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.443442 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-inventory" (OuterVolumeSpecName: "inventory") pod "1202831e-ebbe-4507-b89a-b3e9de9d64e6" (UID: "1202831e-ebbe-4507-b89a-b3e9de9d64e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.483418 4679 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.483463 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.483475 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.483489 4679 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/1202831e-ebbe-4507-b89a-b3e9de9d64e6-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.483502 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghscj\" (UniqueName: \"kubernetes.io/projected/1202831e-ebbe-4507-b89a-b3e9de9d64e6-kube-api-access-ghscj\") on node \"crc\" DevicePath \"\"" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.858023 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" event={"ID":"1202831e-ebbe-4507-b89a-b3e9de9d64e6","Type":"ContainerDied","Data":"eff0d9abae8907883ebb8082dd825ab85483c894e4c88701cc2acc7d40adec12"} Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.858385 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eff0d9abae8907883ebb8082dd825ab85483c894e4c88701cc2acc7d40adec12" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.858070 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lhf76" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.932032 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.932143 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.932221 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.933373 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.933446 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" gracePeriod=600 Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971242 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk"] Dec 02 10:59:16 crc kubenswrapper[4679]: E1202 10:59:16.971672 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="extract-utilities" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971688 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="extract-utilities" Dec 02 10:59:16 crc kubenswrapper[4679]: E1202 10:59:16.971707 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1202831e-ebbe-4507-b89a-b3e9de9d64e6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971715 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="1202831e-ebbe-4507-b89a-b3e9de9d64e6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 10:59:16 crc kubenswrapper[4679]: E1202 10:59:16.971743 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="extract-content" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971749 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="extract-content" Dec 02 10:59:16 crc kubenswrapper[4679]: E1202 10:59:16.971768 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="registry-server" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971775 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="registry-server" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971938 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee66f902-70c0-4953-b6ed-021772b1d0fc" containerName="registry-server" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.971966 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="1202831e-ebbe-4507-b89a-b3e9de9d64e6" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.973561 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.976286 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.976652 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.977271 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.977468 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.977513 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.977621 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.980427 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 10:59:16 crc kubenswrapper[4679]: I1202 10:59:16.988248 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk"] Dec 02 10:59:17 crc kubenswrapper[4679]: E1202 10:59:17.060734 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096184 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcbxd\" (UniqueName: \"kubernetes.io/projected/eae6d050-4b37-4c45-a0bd-4b930852eecf-kube-api-access-xcbxd\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096245 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096434 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096477 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096500 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096592 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096619 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096782 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.096911 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199336 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199426 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199494 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcbxd\" (UniqueName: \"kubernetes.io/projected/eae6d050-4b37-4c45-a0bd-4b930852eecf-kube-api-access-xcbxd\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199525 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199596 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199620 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199643 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199679 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.199703 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.201725 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.204767 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.205076 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.205353 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.207014 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.207933 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.208734 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.209108 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.224420 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcbxd\" (UniqueName: \"kubernetes.io/projected/eae6d050-4b37-4c45-a0bd-4b930852eecf-kube-api-access-xcbxd\") pod \"nova-edpm-deployment-openstack-edpm-ipam-nvlmk\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.308023 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.861721 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk"] Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.874151 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" exitCode=0 Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.874195 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65"} Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.874257 4679 scope.go:117] "RemoveContainer" containerID="9c8b1bf853618b96cd693f0b2fd8cdfb8559d98bac63a4c138f60ed0d408a7d0" Dec 02 10:59:17 crc kubenswrapper[4679]: I1202 10:59:17.874982 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 10:59:17 crc kubenswrapper[4679]: E1202 10:59:17.875249 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:59:18 crc kubenswrapper[4679]: I1202 10:59:18.884464 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" event={"ID":"eae6d050-4b37-4c45-a0bd-4b930852eecf","Type":"ContainerStarted","Data":"07852deec656ed1b5c567971aedfa662a1280078f6c3cd1c762c8609b9add059"} Dec 02 10:59:18 crc kubenswrapper[4679]: I1202 10:59:18.885143 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" event={"ID":"eae6d050-4b37-4c45-a0bd-4b930852eecf","Type":"ContainerStarted","Data":"3eedbd8e5fb8021899a44862d8514698c456761d378f3cd101f454b4adf84537"} Dec 02 10:59:18 crc kubenswrapper[4679]: I1202 10:59:18.913531 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" podStartSLOduration=2.276824337 podStartE2EDuration="2.913496068s" podCreationTimestamp="2025-12-02 10:59:16 +0000 UTC" firstStartedPulling="2025-12-02 10:59:17.864374139 +0000 UTC m=+2411.194512999" lastFinishedPulling="2025-12-02 10:59:18.50104586 +0000 UTC m=+2411.831184730" observedRunningTime="2025-12-02 10:59:18.907316732 +0000 UTC m=+2412.237455592" watchObservedRunningTime="2025-12-02 10:59:18.913496068 +0000 UTC m=+2412.243634938" Dec 02 10:59:32 crc kubenswrapper[4679]: I1202 10:59:32.909700 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 10:59:32 crc kubenswrapper[4679]: E1202 10:59:32.910508 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 10:59:47 crc kubenswrapper[4679]: I1202 10:59:47.909509 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 10:59:47 crc kubenswrapper[4679]: E1202 10:59:47.910182 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.160267 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh"] Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.162297 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.165178 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.168702 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh"] Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.169369 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.307683 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-config-volume\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.307834 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2nhz\" (UniqueName: \"kubernetes.io/projected/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-kube-api-access-g2nhz\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.307924 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-secret-volume\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.409564 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2nhz\" (UniqueName: \"kubernetes.io/projected/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-kube-api-access-g2nhz\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.409621 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-secret-volume\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.409736 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-config-volume\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.410621 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-config-volume\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.417096 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-secret-volume\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.427051 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2nhz\" (UniqueName: \"kubernetes.io/projected/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-kube-api-access-g2nhz\") pod \"collect-profiles-29411220-n52gh\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.481789 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:00 crc kubenswrapper[4679]: I1202 11:00:00.941438 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh"] Dec 02 11:00:00 crc kubenswrapper[4679]: W1202 11:00:00.945005 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c99cf77_95e1_40ff_b7b0_1c7f8575b277.slice/crio-ccdd55a569705311b97d4fa5a81eb74696e04394b418bfd58c1663447c138d7f WatchSource:0}: Error finding container ccdd55a569705311b97d4fa5a81eb74696e04394b418bfd58c1663447c138d7f: Status 404 returned error can't find the container with id ccdd55a569705311b97d4fa5a81eb74696e04394b418bfd58c1663447c138d7f Dec 02 11:00:01 crc kubenswrapper[4679]: I1202 11:00:01.322502 4679 generic.go:334] "Generic (PLEG): container finished" podID="8c99cf77-95e1-40ff-b7b0-1c7f8575b277" containerID="74d61fcea0adb6b8431e1c2e18dbcdb0fba7bf2c9f8119f870a4227f67153916" exitCode=0 Dec 02 11:00:01 crc kubenswrapper[4679]: I1202 11:00:01.322587 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" event={"ID":"8c99cf77-95e1-40ff-b7b0-1c7f8575b277","Type":"ContainerDied","Data":"74d61fcea0adb6b8431e1c2e18dbcdb0fba7bf2c9f8119f870a4227f67153916"} Dec 02 11:00:01 crc kubenswrapper[4679]: I1202 11:00:01.322938 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" event={"ID":"8c99cf77-95e1-40ff-b7b0-1c7f8575b277","Type":"ContainerStarted","Data":"ccdd55a569705311b97d4fa5a81eb74696e04394b418bfd58c1663447c138d7f"} Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.685268 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.874982 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-config-volume\") pod \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.875034 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2nhz\" (UniqueName: \"kubernetes.io/projected/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-kube-api-access-g2nhz\") pod \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.875099 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-secret-volume\") pod \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\" (UID: \"8c99cf77-95e1-40ff-b7b0-1c7f8575b277\") " Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.875973 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-config-volume" (OuterVolumeSpecName: "config-volume") pod "8c99cf77-95e1-40ff-b7b0-1c7f8575b277" (UID: "8c99cf77-95e1-40ff-b7b0-1c7f8575b277"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.882247 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8c99cf77-95e1-40ff-b7b0-1c7f8575b277" (UID: "8c99cf77-95e1-40ff-b7b0-1c7f8575b277"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.883520 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-kube-api-access-g2nhz" (OuterVolumeSpecName: "kube-api-access-g2nhz") pod "8c99cf77-95e1-40ff-b7b0-1c7f8575b277" (UID: "8c99cf77-95e1-40ff-b7b0-1c7f8575b277"). InnerVolumeSpecName "kube-api-access-g2nhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.909622 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:00:02 crc kubenswrapper[4679]: E1202 11:00:02.910204 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.977270 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.977324 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2nhz\" (UniqueName: \"kubernetes.io/projected/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-kube-api-access-g2nhz\") on node \"crc\" DevicePath \"\"" Dec 02 11:00:02 crc kubenswrapper[4679]: I1202 11:00:02.977340 4679 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c99cf77-95e1-40ff-b7b0-1c7f8575b277-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 11:00:03 crc kubenswrapper[4679]: I1202 11:00:03.346670 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" event={"ID":"8c99cf77-95e1-40ff-b7b0-1c7f8575b277","Type":"ContainerDied","Data":"ccdd55a569705311b97d4fa5a81eb74696e04394b418bfd58c1663447c138d7f"} Dec 02 11:00:03 crc kubenswrapper[4679]: I1202 11:00:03.346736 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccdd55a569705311b97d4fa5a81eb74696e04394b418bfd58c1663447c138d7f" Dec 02 11:00:03 crc kubenswrapper[4679]: I1202 11:00:03.346801 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411220-n52gh" Dec 02 11:00:03 crc kubenswrapper[4679]: I1202 11:00:03.783631 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf"] Dec 02 11:00:03 crc kubenswrapper[4679]: I1202 11:00:03.793110 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411175-gkfrf"] Dec 02 11:00:04 crc kubenswrapper[4679]: I1202 11:00:04.920784 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40d1aca-74a1-44e7-aed1-c595ce552803" path="/var/lib/kubelet/pods/d40d1aca-74a1-44e7-aed1-c595ce552803/volumes" Dec 02 11:00:17 crc kubenswrapper[4679]: I1202 11:00:17.908725 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:00:17 crc kubenswrapper[4679]: E1202 11:00:17.909501 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:00:22 crc kubenswrapper[4679]: I1202 11:00:22.961754 4679 scope.go:117] "RemoveContainer" containerID="63bc391a21663da67a2d4db4ae4cf6f7d60cd5513067324208f7e9e29f0a43c8" Dec 02 11:00:29 crc kubenswrapper[4679]: I1202 11:00:29.909969 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:00:29 crc kubenswrapper[4679]: E1202 11:00:29.911058 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:00:42 crc kubenswrapper[4679]: I1202 11:00:42.908887 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:00:42 crc kubenswrapper[4679]: E1202 11:00:42.909851 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:00:56 crc kubenswrapper[4679]: I1202 11:00:56.916539 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:00:56 crc kubenswrapper[4679]: E1202 11:00:56.917505 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.148954 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29411221-9cggb"] Dec 02 11:01:00 crc kubenswrapper[4679]: E1202 11:01:00.151215 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c99cf77-95e1-40ff-b7b0-1c7f8575b277" containerName="collect-profiles" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.151342 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c99cf77-95e1-40ff-b7b0-1c7f8575b277" containerName="collect-profiles" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.151718 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c99cf77-95e1-40ff-b7b0-1c7f8575b277" containerName="collect-profiles" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.152555 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.167106 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29411221-9cggb"] Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.216035 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-combined-ca-bundle\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.216108 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrgmr\" (UniqueName: \"kubernetes.io/projected/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-kube-api-access-nrgmr\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.216525 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-fernet-keys\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.216647 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-config-data\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.318121 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-config-data\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.318235 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-combined-ca-bundle\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.318262 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrgmr\" (UniqueName: \"kubernetes.io/projected/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-kube-api-access-nrgmr\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.318412 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-fernet-keys\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.326176 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-combined-ca-bundle\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.326487 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-fernet-keys\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.327795 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-config-data\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.339673 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrgmr\" (UniqueName: \"kubernetes.io/projected/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-kube-api-access-nrgmr\") pod \"keystone-cron-29411221-9cggb\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.480570 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:00 crc kubenswrapper[4679]: I1202 11:01:00.907036 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29411221-9cggb"] Dec 02 11:01:01 crc kubenswrapper[4679]: I1202 11:01:01.888940 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411221-9cggb" event={"ID":"dea51cbc-b2de-4d00-a13e-fb2e1c66b605","Type":"ContainerStarted","Data":"fbf2fa1a3a711948ce027aecad938c46bd51c598ab4f76f61749f42136cf2c7f"} Dec 02 11:01:01 crc kubenswrapper[4679]: I1202 11:01:01.889398 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411221-9cggb" event={"ID":"dea51cbc-b2de-4d00-a13e-fb2e1c66b605","Type":"ContainerStarted","Data":"e53b59b14822eea44d427b965f04b0dbb7529aa14d7a8bf6476966f9ef9e1998"} Dec 02 11:01:01 crc kubenswrapper[4679]: I1202 11:01:01.912497 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29411221-9cggb" podStartSLOduration=1.912477585 podStartE2EDuration="1.912477585s" podCreationTimestamp="2025-12-02 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 11:01:01.911899228 +0000 UTC m=+2515.242038098" watchObservedRunningTime="2025-12-02 11:01:01.912477585 +0000 UTC m=+2515.242616445" Dec 02 11:01:03 crc kubenswrapper[4679]: I1202 11:01:03.907390 4679 generic.go:334] "Generic (PLEG): container finished" podID="dea51cbc-b2de-4d00-a13e-fb2e1c66b605" containerID="fbf2fa1a3a711948ce027aecad938c46bd51c598ab4f76f61749f42136cf2c7f" exitCode=0 Dec 02 11:01:03 crc kubenswrapper[4679]: I1202 11:01:03.907745 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411221-9cggb" event={"ID":"dea51cbc-b2de-4d00-a13e-fb2e1c66b605","Type":"ContainerDied","Data":"fbf2fa1a3a711948ce027aecad938c46bd51c598ab4f76f61749f42136cf2c7f"} Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.273740 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.370585 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrgmr\" (UniqueName: \"kubernetes.io/projected/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-kube-api-access-nrgmr\") pod \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.370650 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-combined-ca-bundle\") pod \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.370737 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-fernet-keys\") pod \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.370830 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-config-data\") pod \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\" (UID: \"dea51cbc-b2de-4d00-a13e-fb2e1c66b605\") " Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.376556 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-kube-api-access-nrgmr" (OuterVolumeSpecName: "kube-api-access-nrgmr") pod "dea51cbc-b2de-4d00-a13e-fb2e1c66b605" (UID: "dea51cbc-b2de-4d00-a13e-fb2e1c66b605"). InnerVolumeSpecName "kube-api-access-nrgmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.377407 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dea51cbc-b2de-4d00-a13e-fb2e1c66b605" (UID: "dea51cbc-b2de-4d00-a13e-fb2e1c66b605"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.408578 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dea51cbc-b2de-4d00-a13e-fb2e1c66b605" (UID: "dea51cbc-b2de-4d00-a13e-fb2e1c66b605"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.438985 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-config-data" (OuterVolumeSpecName: "config-data") pod "dea51cbc-b2de-4d00-a13e-fb2e1c66b605" (UID: "dea51cbc-b2de-4d00-a13e-fb2e1c66b605"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.473422 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.473462 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrgmr\" (UniqueName: \"kubernetes.io/projected/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-kube-api-access-nrgmr\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.473477 4679 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.473492 4679 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dea51cbc-b2de-4d00-a13e-fb2e1c66b605-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.931091 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411221-9cggb" event={"ID":"dea51cbc-b2de-4d00-a13e-fb2e1c66b605","Type":"ContainerDied","Data":"e53b59b14822eea44d427b965f04b0dbb7529aa14d7a8bf6476966f9ef9e1998"} Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.931429 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e53b59b14822eea44d427b965f04b0dbb7529aa14d7a8bf6476966f9ef9e1998" Dec 02 11:01:05 crc kubenswrapper[4679]: I1202 11:01:05.931138 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411221-9cggb" Dec 02 11:01:08 crc kubenswrapper[4679]: I1202 11:01:08.909405 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:01:08 crc kubenswrapper[4679]: E1202 11:01:08.909881 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:01:23 crc kubenswrapper[4679]: I1202 11:01:23.910098 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:01:23 crc kubenswrapper[4679]: E1202 11:01:23.911525 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:01:35 crc kubenswrapper[4679]: I1202 11:01:35.909208 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:01:35 crc kubenswrapper[4679]: E1202 11:01:35.909946 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:01:50 crc kubenswrapper[4679]: I1202 11:01:50.908930 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:01:50 crc kubenswrapper[4679]: E1202 11:01:50.909700 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:01:57 crc kubenswrapper[4679]: I1202 11:01:57.394260 4679 generic.go:334] "Generic (PLEG): container finished" podID="eae6d050-4b37-4c45-a0bd-4b930852eecf" containerID="07852deec656ed1b5c567971aedfa662a1280078f6c3cd1c762c8609b9add059" exitCode=0 Dec 02 11:01:57 crc kubenswrapper[4679]: I1202 11:01:57.394366 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" event={"ID":"eae6d050-4b37-4c45-a0bd-4b930852eecf","Type":"ContainerDied","Data":"07852deec656ed1b5c567971aedfa662a1280078f6c3cd1c762c8609b9add059"} Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.785690 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.971768 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-ssh-key\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.972198 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-1\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.972853 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcbxd\" (UniqueName: \"kubernetes.io/projected/eae6d050-4b37-4c45-a0bd-4b930852eecf-kube-api-access-xcbxd\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.972948 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-0\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.972990 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-inventory\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.973025 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-1\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.973056 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-combined-ca-bundle\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.973149 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-0\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.973186 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-extra-config-0\") pod \"eae6d050-4b37-4c45-a0bd-4b930852eecf\" (UID: \"eae6d050-4b37-4c45-a0bd-4b930852eecf\") " Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.985614 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:58 crc kubenswrapper[4679]: I1202 11:01:58.986094 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae6d050-4b37-4c45-a0bd-4b930852eecf-kube-api-access-xcbxd" (OuterVolumeSpecName: "kube-api-access-xcbxd") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "kube-api-access-xcbxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.003040 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.006897 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-inventory" (OuterVolumeSpecName: "inventory") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.015622 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.015596 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.016532 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.032651 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.036029 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "eae6d050-4b37-4c45-a0bd-4b930852eecf" (UID: "eae6d050-4b37-4c45-a0bd-4b930852eecf"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082040 4679 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082076 4679 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082085 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082093 4679 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082102 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcbxd\" (UniqueName: \"kubernetes.io/projected/eae6d050-4b37-4c45-a0bd-4b930852eecf-kube-api-access-xcbxd\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082111 4679 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082120 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082128 4679 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.082136 4679 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eae6d050-4b37-4c45-a0bd-4b930852eecf-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.414227 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" event={"ID":"eae6d050-4b37-4c45-a0bd-4b930852eecf","Type":"ContainerDied","Data":"3eedbd8e5fb8021899a44862d8514698c456761d378f3cd101f454b4adf84537"} Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.414267 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3eedbd8e5fb8021899a44862d8514698c456761d378f3cd101f454b4adf84537" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.414347 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-nvlmk" Dec 02 11:01:59 crc kubenswrapper[4679]: E1202 11:01:59.576280 4679 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeae6d050_4b37_4c45_a0bd_4b930852eecf.slice/crio-3eedbd8e5fb8021899a44862d8514698c456761d378f3cd101f454b4adf84537\": RecentStats: unable to find data in memory cache]" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.616932 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6"] Dec 02 11:01:59 crc kubenswrapper[4679]: E1202 11:01:59.628627 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea51cbc-b2de-4d00-a13e-fb2e1c66b605" containerName="keystone-cron" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.628666 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea51cbc-b2de-4d00-a13e-fb2e1c66b605" containerName="keystone-cron" Dec 02 11:01:59 crc kubenswrapper[4679]: E1202 11:01:59.628716 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae6d050-4b37-4c45-a0bd-4b930852eecf" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.628724 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae6d050-4b37-4c45-a0bd-4b930852eecf" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.628988 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae6d050-4b37-4c45-a0bd-4b930852eecf" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.628999 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea51cbc-b2de-4d00-a13e-fb2e1c66b605" containerName="keystone-cron" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.629654 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.635562 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.635612 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.635570 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.635813 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mgmpl" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.635888 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.638251 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6"] Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.702573 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.702918 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.702956 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.703016 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.703036 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6ngl\" (UniqueName: \"kubernetes.io/projected/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-kube-api-access-g6ngl\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.703082 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.703106 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805501 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805603 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805641 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805686 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805713 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6ngl\" (UniqueName: \"kubernetes.io/projected/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-kube-api-access-g6ngl\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805735 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.805767 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.809637 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.809710 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.810379 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.810500 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.811340 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.825052 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.825642 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6ngl\" (UniqueName: \"kubernetes.io/projected/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-kube-api-access-g6ngl\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:01:59 crc kubenswrapper[4679]: I1202 11:01:59.962806 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:02:00 crc kubenswrapper[4679]: I1202 11:02:00.533111 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6"] Dec 02 11:02:01 crc kubenswrapper[4679]: I1202 11:02:01.432552 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" event={"ID":"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8","Type":"ContainerStarted","Data":"81875e9fd54989e826535f227a069ce9dd8d5497023d43a7aace93b403459f3b"} Dec 02 11:02:01 crc kubenswrapper[4679]: I1202 11:02:01.432975 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" event={"ID":"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8","Type":"ContainerStarted","Data":"dd5e82431474ec48587d8a500cf086481af69f317d6bf64ee6eb86cf8a2a74d9"} Dec 02 11:02:02 crc kubenswrapper[4679]: I1202 11:02:02.468520 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" podStartSLOduration=2.819892803 podStartE2EDuration="3.468489543s" podCreationTimestamp="2025-12-02 11:01:59 +0000 UTC" firstStartedPulling="2025-12-02 11:02:00.540749427 +0000 UTC m=+2573.870888287" lastFinishedPulling="2025-12-02 11:02:01.189346167 +0000 UTC m=+2574.519485027" observedRunningTime="2025-12-02 11:02:02.462280876 +0000 UTC m=+2575.792419736" watchObservedRunningTime="2025-12-02 11:02:02.468489543 +0000 UTC m=+2575.798628403" Dec 02 11:02:04 crc kubenswrapper[4679]: I1202 11:02:04.908576 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:02:04 crc kubenswrapper[4679]: E1202 11:02:04.909167 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:02:18 crc kubenswrapper[4679]: I1202 11:02:18.909598 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:02:18 crc kubenswrapper[4679]: E1202 11:02:18.910461 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:02:29 crc kubenswrapper[4679]: I1202 11:02:29.908777 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:02:29 crc kubenswrapper[4679]: E1202 11:02:29.909665 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:02:42 crc kubenswrapper[4679]: I1202 11:02:42.909207 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:02:42 crc kubenswrapper[4679]: E1202 11:02:42.910046 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.805382 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x56fs"] Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.808002 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.818451 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x56fs"] Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.888498 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9lcx\" (UniqueName: \"kubernetes.io/projected/86265308-2402-4516-a3c1-414d1c9859f1-kube-api-access-m9lcx\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.888578 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-utilities\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.888675 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-catalog-content\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.990286 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9lcx\" (UniqueName: \"kubernetes.io/projected/86265308-2402-4516-a3c1-414d1c9859f1-kube-api-access-m9lcx\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.990369 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-utilities\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.990445 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-catalog-content\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.990908 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-catalog-content\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:49 crc kubenswrapper[4679]: I1202 11:02:49.990924 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-utilities\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:50 crc kubenswrapper[4679]: I1202 11:02:50.024380 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9lcx\" (UniqueName: \"kubernetes.io/projected/86265308-2402-4516-a3c1-414d1c9859f1-kube-api-access-m9lcx\") pod \"redhat-operators-x56fs\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:50 crc kubenswrapper[4679]: I1202 11:02:50.146407 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:02:50 crc kubenswrapper[4679]: I1202 11:02:50.613860 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x56fs"] Dec 02 11:02:50 crc kubenswrapper[4679]: I1202 11:02:50.853621 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerStarted","Data":"6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4"} Dec 02 11:02:50 crc kubenswrapper[4679]: I1202 11:02:50.853941 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerStarted","Data":"f6c011ea15afdd6737b05cc00dd30015fe3ae4a698278ff0a1154705024dc2dc"} Dec 02 11:02:50 crc kubenswrapper[4679]: I1202 11:02:50.858983 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 11:02:51 crc kubenswrapper[4679]: I1202 11:02:51.872098 4679 generic.go:334] "Generic (PLEG): container finished" podID="86265308-2402-4516-a3c1-414d1c9859f1" containerID="6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4" exitCode=0 Dec 02 11:02:51 crc kubenswrapper[4679]: I1202 11:02:51.872230 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerDied","Data":"6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4"} Dec 02 11:02:52 crc kubenswrapper[4679]: I1202 11:02:52.883151 4679 generic.go:334] "Generic (PLEG): container finished" podID="86265308-2402-4516-a3c1-414d1c9859f1" containerID="c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42" exitCode=0 Dec 02 11:02:52 crc kubenswrapper[4679]: I1202 11:02:52.883197 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerDied","Data":"c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42"} Dec 02 11:02:54 crc kubenswrapper[4679]: I1202 11:02:54.918121 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerStarted","Data":"37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390"} Dec 02 11:02:54 crc kubenswrapper[4679]: I1202 11:02:54.932023 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x56fs" podStartSLOduration=2.743636744 podStartE2EDuration="5.932001499s" podCreationTimestamp="2025-12-02 11:02:49 +0000 UTC" firstStartedPulling="2025-12-02 11:02:50.858720478 +0000 UTC m=+2624.188859338" lastFinishedPulling="2025-12-02 11:02:54.047085233 +0000 UTC m=+2627.377224093" observedRunningTime="2025-12-02 11:02:54.930220299 +0000 UTC m=+2628.260359159" watchObservedRunningTime="2025-12-02 11:02:54.932001499 +0000 UTC m=+2628.262140359" Dec 02 11:02:57 crc kubenswrapper[4679]: I1202 11:02:57.909432 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:02:57 crc kubenswrapper[4679]: E1202 11:02:57.910015 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:03:00 crc kubenswrapper[4679]: I1202 11:03:00.146702 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:03:00 crc kubenswrapper[4679]: I1202 11:03:00.147901 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:03:00 crc kubenswrapper[4679]: I1202 11:03:00.249516 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:03:01 crc kubenswrapper[4679]: I1202 11:03:01.004052 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:03:01 crc kubenswrapper[4679]: I1202 11:03:01.068532 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x56fs"] Dec 02 11:03:02 crc kubenswrapper[4679]: I1202 11:03:02.978888 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x56fs" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="registry-server" containerID="cri-o://37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390" gracePeriod=2 Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.435934 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.570015 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-catalog-content\") pod \"86265308-2402-4516-a3c1-414d1c9859f1\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.570170 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9lcx\" (UniqueName: \"kubernetes.io/projected/86265308-2402-4516-a3c1-414d1c9859f1-kube-api-access-m9lcx\") pod \"86265308-2402-4516-a3c1-414d1c9859f1\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.570425 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-utilities\") pod \"86265308-2402-4516-a3c1-414d1c9859f1\" (UID: \"86265308-2402-4516-a3c1-414d1c9859f1\") " Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.572462 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-utilities" (OuterVolumeSpecName: "utilities") pod "86265308-2402-4516-a3c1-414d1c9859f1" (UID: "86265308-2402-4516-a3c1-414d1c9859f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.581944 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86265308-2402-4516-a3c1-414d1c9859f1-kube-api-access-m9lcx" (OuterVolumeSpecName: "kube-api-access-m9lcx") pod "86265308-2402-4516-a3c1-414d1c9859f1" (UID: "86265308-2402-4516-a3c1-414d1c9859f1"). InnerVolumeSpecName "kube-api-access-m9lcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.673070 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9lcx\" (UniqueName: \"kubernetes.io/projected/86265308-2402-4516-a3c1-414d1c9859f1-kube-api-access-m9lcx\") on node \"crc\" DevicePath \"\"" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.673371 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.693750 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86265308-2402-4516-a3c1-414d1c9859f1" (UID: "86265308-2402-4516-a3c1-414d1c9859f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.775042 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86265308-2402-4516-a3c1-414d1c9859f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.994752 4679 generic.go:334] "Generic (PLEG): container finished" podID="86265308-2402-4516-a3c1-414d1c9859f1" containerID="37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390" exitCode=0 Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.994793 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerDied","Data":"37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390"} Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.994820 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56fs" event={"ID":"86265308-2402-4516-a3c1-414d1c9859f1","Type":"ContainerDied","Data":"f6c011ea15afdd6737b05cc00dd30015fe3ae4a698278ff0a1154705024dc2dc"} Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.994839 4679 scope.go:117] "RemoveContainer" containerID="37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390" Dec 02 11:03:03 crc kubenswrapper[4679]: I1202 11:03:03.994957 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56fs" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.033923 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x56fs"] Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.042836 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x56fs"] Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.061519 4679 scope.go:117] "RemoveContainer" containerID="c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.086988 4679 scope.go:117] "RemoveContainer" containerID="6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.147711 4679 scope.go:117] "RemoveContainer" containerID="37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390" Dec 02 11:03:04 crc kubenswrapper[4679]: E1202 11:03:04.148262 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390\": container with ID starting with 37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390 not found: ID does not exist" containerID="37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.148320 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390"} err="failed to get container status \"37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390\": rpc error: code = NotFound desc = could not find container \"37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390\": container with ID starting with 37f52449a1199c487fd09167c9804b2926e91685be069b7d556f03ce9c1ae390 not found: ID does not exist" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.148347 4679 scope.go:117] "RemoveContainer" containerID="c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42" Dec 02 11:03:04 crc kubenswrapper[4679]: E1202 11:03:04.148959 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42\": container with ID starting with c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42 not found: ID does not exist" containerID="c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.149015 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42"} err="failed to get container status \"c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42\": rpc error: code = NotFound desc = could not find container \"c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42\": container with ID starting with c7fea353abf98d7faf546b558195b4349cc628f9cf0f1ecd708d3e59c628ce42 not found: ID does not exist" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.149052 4679 scope.go:117] "RemoveContainer" containerID="6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4" Dec 02 11:03:04 crc kubenswrapper[4679]: E1202 11:03:04.149428 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4\": container with ID starting with 6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4 not found: ID does not exist" containerID="6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.149459 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4"} err="failed to get container status \"6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4\": rpc error: code = NotFound desc = could not find container \"6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4\": container with ID starting with 6a1f8305d4ecd4859a70231544933c0ea0acf1d1bf6d3ef45b4a048158a8d9b4 not found: ID does not exist" Dec 02 11:03:04 crc kubenswrapper[4679]: I1202 11:03:04.929856 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86265308-2402-4516-a3c1-414d1c9859f1" path="/var/lib/kubelet/pods/86265308-2402-4516-a3c1-414d1c9859f1/volumes" Dec 02 11:03:11 crc kubenswrapper[4679]: I1202 11:03:11.908326 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:03:11 crc kubenswrapper[4679]: E1202 11:03:11.909285 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:03:22 crc kubenswrapper[4679]: I1202 11:03:22.909437 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:03:22 crc kubenswrapper[4679]: E1202 11:03:22.910345 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:03:37 crc kubenswrapper[4679]: I1202 11:03:37.909364 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:03:37 crc kubenswrapper[4679]: E1202 11:03:37.910544 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:03:52 crc kubenswrapper[4679]: I1202 11:03:52.908875 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:03:52 crc kubenswrapper[4679]: E1202 11:03:52.909461 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:04:07 crc kubenswrapper[4679]: I1202 11:04:07.910008 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:04:07 crc kubenswrapper[4679]: E1202 11:04:07.910878 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:04:12 crc kubenswrapper[4679]: I1202 11:04:12.609470 4679 generic.go:334] "Generic (PLEG): container finished" podID="d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" containerID="81875e9fd54989e826535f227a069ce9dd8d5497023d43a7aace93b403459f3b" exitCode=0 Dec 02 11:04:12 crc kubenswrapper[4679]: I1202 11:04:12.609572 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" event={"ID":"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8","Type":"ContainerDied","Data":"81875e9fd54989e826535f227a069ce9dd8d5497023d43a7aace93b403459f3b"} Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.083147 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238105 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-2\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238277 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6ngl\" (UniqueName: \"kubernetes.io/projected/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-kube-api-access-g6ngl\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238503 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ssh-key\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238557 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-inventory\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238594 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-telemetry-combined-ca-bundle\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238625 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-0\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.238651 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-1\") pod \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\" (UID: \"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8\") " Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.243999 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-kube-api-access-g6ngl" (OuterVolumeSpecName: "kube-api-access-g6ngl") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "kube-api-access-g6ngl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.247622 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.266395 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.267434 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-inventory" (OuterVolumeSpecName: "inventory") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.269485 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.269734 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.271195 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" (UID: "d79ba6bd-2e41-485d-a11f-5b0bdb5500c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341251 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341322 4679 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341336 4679 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341348 4679 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341358 4679 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341367 4679 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.341378 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6ngl\" (UniqueName: \"kubernetes.io/projected/d79ba6bd-2e41-485d-a11f-5b0bdb5500c8-kube-api-access-g6ngl\") on node \"crc\" DevicePath \"\"" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.638005 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" event={"ID":"d79ba6bd-2e41-485d-a11f-5b0bdb5500c8","Type":"ContainerDied","Data":"dd5e82431474ec48587d8a500cf086481af69f317d6bf64ee6eb86cf8a2a74d9"} Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.638055 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd5e82431474ec48587d8a500cf086481af69f317d6bf64ee6eb86cf8a2a74d9" Dec 02 11:04:14 crc kubenswrapper[4679]: I1202 11:04:14.638158 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6" Dec 02 11:04:22 crc kubenswrapper[4679]: I1202 11:04:22.909089 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:04:23 crc kubenswrapper[4679]: I1202 11:04:23.726140 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"70127d6ce84a03c1c8b8d09d74c1b2753f1cfd4c549cf4f0d5c6e12b15b2f314"} Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.708731 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 11:05:16 crc kubenswrapper[4679]: E1202 11:05:16.710531 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="extract-content" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.710620 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="extract-content" Dec 02 11:05:16 crc kubenswrapper[4679]: E1202 11:05:16.710697 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="extract-utilities" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.710750 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="extract-utilities" Dec 02 11:05:16 crc kubenswrapper[4679]: E1202 11:05:16.710828 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.710908 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 11:05:16 crc kubenswrapper[4679]: E1202 11:05:16.711009 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="registry-server" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.711086 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="registry-server" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.711415 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="86265308-2402-4516-a3c1-414d1c9859f1" containerName="registry-server" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.711534 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79ba6bd-2e41-485d-a11f-5b0bdb5500c8" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.712457 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.715721 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.715721 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.720409 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.721186 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5xprf" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.735131 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.811481 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-config-data\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.811733 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vslz\" (UniqueName: \"kubernetes.io/projected/564a5130-acc1-40e0-814b-2e9bfd2d967d-kube-api-access-9vslz\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.811843 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.811998 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.812075 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.812194 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.812268 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.812399 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.812483 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914365 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-config-data\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914417 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vslz\" (UniqueName: \"kubernetes.io/projected/564a5130-acc1-40e0-814b-2e9bfd2d967d-kube-api-access-9vslz\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914446 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914560 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914588 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914632 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914657 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914693 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.914707 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.915145 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.915745 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-config-data\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.917333 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.917607 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.918539 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.922733 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.923498 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.923611 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.933869 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vslz\" (UniqueName: \"kubernetes.io/projected/564a5130-acc1-40e0-814b-2e9bfd2d967d-kube-api-access-9vslz\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:16 crc kubenswrapper[4679]: I1202 11:05:16.959203 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " pod="openstack/tempest-tests-tempest" Dec 02 11:05:17 crc kubenswrapper[4679]: I1202 11:05:17.056125 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 11:05:17 crc kubenswrapper[4679]: I1202 11:05:17.521456 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 11:05:18 crc kubenswrapper[4679]: I1202 11:05:18.239678 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564a5130-acc1-40e0-814b-2e9bfd2d967d","Type":"ContainerStarted","Data":"39f51430602728ed632223b28025488d5a4e060594fd9447b0b4a890349f37a3"} Dec 02 11:05:51 crc kubenswrapper[4679]: E1202 11:05:51.183429 4679 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 02 11:05:51 crc kubenswrapper[4679]: E1202 11:05:51.184211 4679 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vslz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(564a5130-acc1-40e0-814b-2e9bfd2d967d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 11:05:51 crc kubenswrapper[4679]: E1202 11:05:51.185572 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="564a5130-acc1-40e0-814b-2e9bfd2d967d" Dec 02 11:05:51 crc kubenswrapper[4679]: E1202 11:05:51.553432 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="564a5130-acc1-40e0-814b-2e9bfd2d967d" Dec 02 11:06:07 crc kubenswrapper[4679]: I1202 11:06:07.708468 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564a5130-acc1-40e0-814b-2e9bfd2d967d","Type":"ContainerStarted","Data":"8914a691fce4dac0dd056eef6fe9098a82781955baca29357adfc15601292d2c"} Dec 02 11:06:07 crc kubenswrapper[4679]: I1202 11:06:07.752013 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.798416108 podStartE2EDuration="52.751981948s" podCreationTimestamp="2025-12-02 11:05:15 +0000 UTC" firstStartedPulling="2025-12-02 11:05:17.534615721 +0000 UTC m=+2770.864754591" lastFinishedPulling="2025-12-02 11:06:06.488181571 +0000 UTC m=+2819.818320431" observedRunningTime="2025-12-02 11:06:07.73479547 +0000 UTC m=+2821.064934400" watchObservedRunningTime="2025-12-02 11:06:07.751981948 +0000 UTC m=+2821.082120848" Dec 02 11:06:46 crc kubenswrapper[4679]: I1202 11:06:46.931446 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:06:46 crc kubenswrapper[4679]: I1202 11:06:46.931896 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:07:16 crc kubenswrapper[4679]: I1202 11:07:16.932189 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:07:16 crc kubenswrapper[4679]: I1202 11:07:16.932917 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:07:46 crc kubenswrapper[4679]: I1202 11:07:46.931530 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:07:46 crc kubenswrapper[4679]: I1202 11:07:46.932121 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:07:46 crc kubenswrapper[4679]: I1202 11:07:46.932163 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 11:07:46 crc kubenswrapper[4679]: I1202 11:07:46.932951 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"70127d6ce84a03c1c8b8d09d74c1b2753f1cfd4c549cf4f0d5c6e12b15b2f314"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 11:07:46 crc kubenswrapper[4679]: I1202 11:07:46.933004 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://70127d6ce84a03c1c8b8d09d74c1b2753f1cfd4c549cf4f0d5c6e12b15b2f314" gracePeriod=600 Dec 02 11:07:47 crc kubenswrapper[4679]: I1202 11:07:47.680841 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="70127d6ce84a03c1c8b8d09d74c1b2753f1cfd4c549cf4f0d5c6e12b15b2f314" exitCode=0 Dec 02 11:07:47 crc kubenswrapper[4679]: I1202 11:07:47.681007 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"70127d6ce84a03c1c8b8d09d74c1b2753f1cfd4c549cf4f0d5c6e12b15b2f314"} Dec 02 11:07:47 crc kubenswrapper[4679]: I1202 11:07:47.681219 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a"} Dec 02 11:07:47 crc kubenswrapper[4679]: I1202 11:07:47.681247 4679 scope.go:117] "RemoveContainer" containerID="07f6e8b2040232b278c8a271312d23d9227bfddd4249a63d3ba80892c3661b65" Dec 02 11:08:20 crc kubenswrapper[4679]: I1202 11:08:20.977235 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dpk2b"] Dec 02 11:08:20 crc kubenswrapper[4679]: I1202 11:08:20.979715 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:20 crc kubenswrapper[4679]: I1202 11:08:20.996705 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dpk2b"] Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.149405 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5xd9\" (UniqueName: \"kubernetes.io/projected/f0938f55-510d-4ed1-b71a-05d48e6fd14c-kube-api-access-c5xd9\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.149452 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-utilities\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.149471 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-catalog-content\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.251867 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5xd9\" (UniqueName: \"kubernetes.io/projected/f0938f55-510d-4ed1-b71a-05d48e6fd14c-kube-api-access-c5xd9\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.252738 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-utilities\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.252785 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-utilities\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.252832 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-catalog-content\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.253154 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-catalog-content\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.278582 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5xd9\" (UniqueName: \"kubernetes.io/projected/f0938f55-510d-4ed1-b71a-05d48e6fd14c-kube-api-access-c5xd9\") pod \"community-operators-dpk2b\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.314619 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.879027 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dpk2b"] Dec 02 11:08:21 crc kubenswrapper[4679]: I1202 11:08:21.994623 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpk2b" event={"ID":"f0938f55-510d-4ed1-b71a-05d48e6fd14c","Type":"ContainerStarted","Data":"a7df20431f16d588a5a455484fdcdfdbe026a33a5dc71f3d5655dbb3a7d99c95"} Dec 02 11:08:23 crc kubenswrapper[4679]: I1202 11:08:23.004640 4679 generic.go:334] "Generic (PLEG): container finished" podID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerID="6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1" exitCode=0 Dec 02 11:08:23 crc kubenswrapper[4679]: I1202 11:08:23.004685 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpk2b" event={"ID":"f0938f55-510d-4ed1-b71a-05d48e6fd14c","Type":"ContainerDied","Data":"6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1"} Dec 02 11:08:23 crc kubenswrapper[4679]: I1202 11:08:23.007378 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 11:08:26 crc kubenswrapper[4679]: I1202 11:08:26.036744 4679 generic.go:334] "Generic (PLEG): container finished" podID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerID="cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929" exitCode=0 Dec 02 11:08:26 crc kubenswrapper[4679]: I1202 11:08:26.036863 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpk2b" event={"ID":"f0938f55-510d-4ed1-b71a-05d48e6fd14c","Type":"ContainerDied","Data":"cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929"} Dec 02 11:08:27 crc kubenswrapper[4679]: I1202 11:08:27.047109 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpk2b" event={"ID":"f0938f55-510d-4ed1-b71a-05d48e6fd14c","Type":"ContainerStarted","Data":"9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871"} Dec 02 11:08:27 crc kubenswrapper[4679]: I1202 11:08:27.072023 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dpk2b" podStartSLOduration=3.5656094940000003 podStartE2EDuration="7.072006009s" podCreationTimestamp="2025-12-02 11:08:20 +0000 UTC" firstStartedPulling="2025-12-02 11:08:23.007142466 +0000 UTC m=+2956.337281326" lastFinishedPulling="2025-12-02 11:08:26.513538971 +0000 UTC m=+2959.843677841" observedRunningTime="2025-12-02 11:08:27.062786098 +0000 UTC m=+2960.392924958" watchObservedRunningTime="2025-12-02 11:08:27.072006009 +0000 UTC m=+2960.402144869" Dec 02 11:08:31 crc kubenswrapper[4679]: I1202 11:08:31.315132 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:31 crc kubenswrapper[4679]: I1202 11:08:31.315693 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:31 crc kubenswrapper[4679]: I1202 11:08:31.359387 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:31 crc kubenswrapper[4679]: I1202 11:08:31.983885 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s4drc"] Dec 02 11:08:31 crc kubenswrapper[4679]: I1202 11:08:31.986840 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:31 crc kubenswrapper[4679]: I1202 11:08:31.997778 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s4drc"] Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.144536 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.168505 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-catalog-content\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.168675 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhnkp\" (UniqueName: \"kubernetes.io/projected/3bfda4e3-b65e-40b4-93f8-c7891a983cce-kube-api-access-zhnkp\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.168986 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-utilities\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.270337 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-utilities\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.270408 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-catalog-content\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.270497 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhnkp\" (UniqueName: \"kubernetes.io/projected/3bfda4e3-b65e-40b4-93f8-c7891a983cce-kube-api-access-zhnkp\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.270879 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-utilities\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.271052 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-catalog-content\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.291097 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhnkp\" (UniqueName: \"kubernetes.io/projected/3bfda4e3-b65e-40b4-93f8-c7891a983cce-kube-api-access-zhnkp\") pod \"certified-operators-s4drc\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.313196 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:32 crc kubenswrapper[4679]: I1202 11:08:32.830489 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s4drc"] Dec 02 11:08:33 crc kubenswrapper[4679]: I1202 11:08:33.104231 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4drc" event={"ID":"3bfda4e3-b65e-40b4-93f8-c7891a983cce","Type":"ContainerStarted","Data":"4ed38b23b175c42e94527d23648e09a20edddf4f72f7e10a93bc4e306ee9f0bd"} Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.114422 4679 generic.go:334] "Generic (PLEG): container finished" podID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerID="ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a" exitCode=0 Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.114461 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4drc" event={"ID":"3bfda4e3-b65e-40b4-93f8-c7891a983cce","Type":"ContainerDied","Data":"ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a"} Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.592169 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v82j9"] Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.595660 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.618423 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v82j9"] Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.727575 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-catalog-content\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.727667 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-utilities\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.727839 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r86m\" (UniqueName: \"kubernetes.io/projected/3072c031-d396-444d-b7b6-b1085b54b1f1-kube-api-access-5r86m\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.830416 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r86m\" (UniqueName: \"kubernetes.io/projected/3072c031-d396-444d-b7b6-b1085b54b1f1-kube-api-access-5r86m\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.830814 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-catalog-content\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.830855 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-utilities\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.831465 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-utilities\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.831476 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-catalog-content\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.854578 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r86m\" (UniqueName: \"kubernetes.io/projected/3072c031-d396-444d-b7b6-b1085b54b1f1-kube-api-access-5r86m\") pod \"redhat-marketplace-v82j9\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:34 crc kubenswrapper[4679]: I1202 11:08:34.921769 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:35 crc kubenswrapper[4679]: I1202 11:08:35.369828 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v82j9"] Dec 02 11:08:35 crc kubenswrapper[4679]: W1202 11:08:35.412350 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3072c031_d396_444d_b7b6_b1085b54b1f1.slice/crio-92a1edfee7342b2898f80c99839cd252fea48869c295a4253d82d646c807395e WatchSource:0}: Error finding container 92a1edfee7342b2898f80c99839cd252fea48869c295a4253d82d646c807395e: Status 404 returned error can't find the container with id 92a1edfee7342b2898f80c99839cd252fea48869c295a4253d82d646c807395e Dec 02 11:08:36 crc kubenswrapper[4679]: I1202 11:08:36.140182 4679 generic.go:334] "Generic (PLEG): container finished" podID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerID="0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa" exitCode=0 Dec 02 11:08:36 crc kubenswrapper[4679]: I1202 11:08:36.140238 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v82j9" event={"ID":"3072c031-d396-444d-b7b6-b1085b54b1f1","Type":"ContainerDied","Data":"0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa"} Dec 02 11:08:36 crc kubenswrapper[4679]: I1202 11:08:36.140263 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v82j9" event={"ID":"3072c031-d396-444d-b7b6-b1085b54b1f1","Type":"ContainerStarted","Data":"92a1edfee7342b2898f80c99839cd252fea48869c295a4253d82d646c807395e"} Dec 02 11:08:36 crc kubenswrapper[4679]: I1202 11:08:36.142091 4679 generic.go:334] "Generic (PLEG): container finished" podID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerID="1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c" exitCode=0 Dec 02 11:08:36 crc kubenswrapper[4679]: I1202 11:08:36.142116 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4drc" event={"ID":"3bfda4e3-b65e-40b4-93f8-c7891a983cce","Type":"ContainerDied","Data":"1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c"} Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.153855 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4drc" event={"ID":"3bfda4e3-b65e-40b4-93f8-c7891a983cce","Type":"ContainerStarted","Data":"e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba"} Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.168534 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dpk2b"] Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.168883 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dpk2b" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="registry-server" containerID="cri-o://9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871" gracePeriod=2 Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.187938 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s4drc" podStartSLOduration=3.701136943 podStartE2EDuration="6.187920245s" podCreationTimestamp="2025-12-02 11:08:31 +0000 UTC" firstStartedPulling="2025-12-02 11:08:34.117070158 +0000 UTC m=+2967.447209018" lastFinishedPulling="2025-12-02 11:08:36.60385346 +0000 UTC m=+2969.933992320" observedRunningTime="2025-12-02 11:08:37.180865465 +0000 UTC m=+2970.511004325" watchObservedRunningTime="2025-12-02 11:08:37.187920245 +0000 UTC m=+2970.518059105" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.665330 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.787124 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5xd9\" (UniqueName: \"kubernetes.io/projected/f0938f55-510d-4ed1-b71a-05d48e6fd14c-kube-api-access-c5xd9\") pod \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.787244 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-catalog-content\") pod \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.787344 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-utilities\") pod \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\" (UID: \"f0938f55-510d-4ed1-b71a-05d48e6fd14c\") " Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.787979 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-utilities" (OuterVolumeSpecName: "utilities") pod "f0938f55-510d-4ed1-b71a-05d48e6fd14c" (UID: "f0938f55-510d-4ed1-b71a-05d48e6fd14c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.793855 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0938f55-510d-4ed1-b71a-05d48e6fd14c-kube-api-access-c5xd9" (OuterVolumeSpecName: "kube-api-access-c5xd9") pod "f0938f55-510d-4ed1-b71a-05d48e6fd14c" (UID: "f0938f55-510d-4ed1-b71a-05d48e6fd14c"). InnerVolumeSpecName "kube-api-access-c5xd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.833223 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0938f55-510d-4ed1-b71a-05d48e6fd14c" (UID: "f0938f55-510d-4ed1-b71a-05d48e6fd14c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.890206 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.890244 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5xd9\" (UniqueName: \"kubernetes.io/projected/f0938f55-510d-4ed1-b71a-05d48e6fd14c-kube-api-access-c5xd9\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:37 crc kubenswrapper[4679]: I1202 11:08:37.890261 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0938f55-510d-4ed1-b71a-05d48e6fd14c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.166417 4679 generic.go:334] "Generic (PLEG): container finished" podID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerID="9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871" exitCode=0 Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.166459 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpk2b" event={"ID":"f0938f55-510d-4ed1-b71a-05d48e6fd14c","Type":"ContainerDied","Data":"9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871"} Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.166501 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpk2b" event={"ID":"f0938f55-510d-4ed1-b71a-05d48e6fd14c","Type":"ContainerDied","Data":"a7df20431f16d588a5a455484fdcdfdbe026a33a5dc71f3d5655dbb3a7d99c95"} Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.166519 4679 scope.go:117] "RemoveContainer" containerID="9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.166517 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpk2b" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.168473 4679 generic.go:334] "Generic (PLEG): container finished" podID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerID="e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29" exitCode=0 Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.168515 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v82j9" event={"ID":"3072c031-d396-444d-b7b6-b1085b54b1f1","Type":"ContainerDied","Data":"e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29"} Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.190716 4679 scope.go:117] "RemoveContainer" containerID="cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.228475 4679 scope.go:117] "RemoveContainer" containerID="6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.235039 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dpk2b"] Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.244176 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dpk2b"] Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.260967 4679 scope.go:117] "RemoveContainer" containerID="9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871" Dec 02 11:08:38 crc kubenswrapper[4679]: E1202 11:08:38.261553 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871\": container with ID starting with 9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871 not found: ID does not exist" containerID="9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.261613 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871"} err="failed to get container status \"9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871\": rpc error: code = NotFound desc = could not find container \"9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871\": container with ID starting with 9379a5f9c37b6176b8f3228319be112967aa8f1b399294d24b9365863b463871 not found: ID does not exist" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.261641 4679 scope.go:117] "RemoveContainer" containerID="cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929" Dec 02 11:08:38 crc kubenswrapper[4679]: E1202 11:08:38.262380 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929\": container with ID starting with cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929 not found: ID does not exist" containerID="cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.262418 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929"} err="failed to get container status \"cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929\": rpc error: code = NotFound desc = could not find container \"cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929\": container with ID starting with cadf0125f941ca373e776c3e316d284e000ceeb1279b00b64a6a7d4901812929 not found: ID does not exist" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.262449 4679 scope.go:117] "RemoveContainer" containerID="6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1" Dec 02 11:08:38 crc kubenswrapper[4679]: E1202 11:08:38.262701 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1\": container with ID starting with 6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1 not found: ID does not exist" containerID="6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.262792 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1"} err="failed to get container status \"6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1\": rpc error: code = NotFound desc = could not find container \"6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1\": container with ID starting with 6b3ed0b48dbd7dc44774f9df1ce2cd64514c3fac0403aa2e162924e2efd99eb1 not found: ID does not exist" Dec 02 11:08:38 crc kubenswrapper[4679]: I1202 11:08:38.923273 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" path="/var/lib/kubelet/pods/f0938f55-510d-4ed1-b71a-05d48e6fd14c/volumes" Dec 02 11:08:39 crc kubenswrapper[4679]: I1202 11:08:39.188398 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v82j9" event={"ID":"3072c031-d396-444d-b7b6-b1085b54b1f1","Type":"ContainerStarted","Data":"e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc"} Dec 02 11:08:39 crc kubenswrapper[4679]: I1202 11:08:39.212976 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v82j9" podStartSLOduration=2.672705485 podStartE2EDuration="5.212955676s" podCreationTimestamp="2025-12-02 11:08:34 +0000 UTC" firstStartedPulling="2025-12-02 11:08:36.142018156 +0000 UTC m=+2969.472157016" lastFinishedPulling="2025-12-02 11:08:38.682268337 +0000 UTC m=+2972.012407207" observedRunningTime="2025-12-02 11:08:39.209997332 +0000 UTC m=+2972.540136192" watchObservedRunningTime="2025-12-02 11:08:39.212955676 +0000 UTC m=+2972.543094536" Dec 02 11:08:42 crc kubenswrapper[4679]: I1202 11:08:42.314021 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:42 crc kubenswrapper[4679]: I1202 11:08:42.314387 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:42 crc kubenswrapper[4679]: I1202 11:08:42.378982 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:43 crc kubenswrapper[4679]: I1202 11:08:43.283479 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:43 crc kubenswrapper[4679]: I1202 11:08:43.568698 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s4drc"] Dec 02 11:08:44 crc kubenswrapper[4679]: I1202 11:08:44.921976 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:44 crc kubenswrapper[4679]: I1202 11:08:44.922240 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:44 crc kubenswrapper[4679]: I1202 11:08:44.967422 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.249816 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s4drc" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="registry-server" containerID="cri-o://e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba" gracePeriod=2 Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.297013 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.751759 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.757684 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-utilities\") pod \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.757908 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhnkp\" (UniqueName: \"kubernetes.io/projected/3bfda4e3-b65e-40b4-93f8-c7891a983cce-kube-api-access-zhnkp\") pod \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.757955 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-catalog-content\") pod \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\" (UID: \"3bfda4e3-b65e-40b4-93f8-c7891a983cce\") " Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.758723 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-utilities" (OuterVolumeSpecName: "utilities") pod "3bfda4e3-b65e-40b4-93f8-c7891a983cce" (UID: "3bfda4e3-b65e-40b4-93f8-c7891a983cce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.764892 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bfda4e3-b65e-40b4-93f8-c7891a983cce-kube-api-access-zhnkp" (OuterVolumeSpecName: "kube-api-access-zhnkp") pod "3bfda4e3-b65e-40b4-93f8-c7891a983cce" (UID: "3bfda4e3-b65e-40b4-93f8-c7891a983cce"). InnerVolumeSpecName "kube-api-access-zhnkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.860059 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.860105 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhnkp\" (UniqueName: \"kubernetes.io/projected/3bfda4e3-b65e-40b4-93f8-c7891a983cce-kube-api-access-zhnkp\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:45 crc kubenswrapper[4679]: I1202 11:08:45.995230 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bfda4e3-b65e-40b4-93f8-c7891a983cce" (UID: "3bfda4e3-b65e-40b4-93f8-c7891a983cce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.063074 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bfda4e3-b65e-40b4-93f8-c7891a983cce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.267666 4679 generic.go:334] "Generic (PLEG): container finished" podID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerID="e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba" exitCode=0 Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.267774 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4drc" event={"ID":"3bfda4e3-b65e-40b4-93f8-c7891a983cce","Type":"ContainerDied","Data":"e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba"} Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.267835 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4drc" event={"ID":"3bfda4e3-b65e-40b4-93f8-c7891a983cce","Type":"ContainerDied","Data":"4ed38b23b175c42e94527d23648e09a20edddf4f72f7e10a93bc4e306ee9f0bd"} Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.267845 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4drc" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.267863 4679 scope.go:117] "RemoveContainer" containerID="e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.318419 4679 scope.go:117] "RemoveContainer" containerID="1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.322749 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s4drc"] Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.333070 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s4drc"] Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.342952 4679 scope.go:117] "RemoveContainer" containerID="ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.384554 4679 scope.go:117] "RemoveContainer" containerID="e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba" Dec 02 11:08:46 crc kubenswrapper[4679]: E1202 11:08:46.385056 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba\": container with ID starting with e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba not found: ID does not exist" containerID="e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.385111 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba"} err="failed to get container status \"e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba\": rpc error: code = NotFound desc = could not find container \"e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba\": container with ID starting with e934b17accec48caaa570f4e866a32397e79873af84bb3d5ca5c5c726c8f1bba not found: ID does not exist" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.385144 4679 scope.go:117] "RemoveContainer" containerID="1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c" Dec 02 11:08:46 crc kubenswrapper[4679]: E1202 11:08:46.385583 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c\": container with ID starting with 1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c not found: ID does not exist" containerID="1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.385608 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c"} err="failed to get container status \"1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c\": rpc error: code = NotFound desc = could not find container \"1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c\": container with ID starting with 1708e4701b09a9eb5fb4dfd944884dcf199f974644410116e42d392046c7608c not found: ID does not exist" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.385620 4679 scope.go:117] "RemoveContainer" containerID="ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a" Dec 02 11:08:46 crc kubenswrapper[4679]: E1202 11:08:46.385893 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a\": container with ID starting with ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a not found: ID does not exist" containerID="ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.385915 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a"} err="failed to get container status \"ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a\": rpc error: code = NotFound desc = could not find container \"ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a\": container with ID starting with ffa0501faac12fb39319d8a468651ce02629e7b501bc2b48fe270c609482861a not found: ID does not exist" Dec 02 11:08:46 crc kubenswrapper[4679]: I1202 11:08:46.923792 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" path="/var/lib/kubelet/pods/3bfda4e3-b65e-40b4-93f8-c7891a983cce/volumes" Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.367576 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v82j9"] Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.367852 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v82j9" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="registry-server" containerID="cri-o://e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc" gracePeriod=2 Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.875996 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.998246 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-utilities\") pod \"3072c031-d396-444d-b7b6-b1085b54b1f1\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.998497 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-catalog-content\") pod \"3072c031-d396-444d-b7b6-b1085b54b1f1\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.998551 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r86m\" (UniqueName: \"kubernetes.io/projected/3072c031-d396-444d-b7b6-b1085b54b1f1-kube-api-access-5r86m\") pod \"3072c031-d396-444d-b7b6-b1085b54b1f1\" (UID: \"3072c031-d396-444d-b7b6-b1085b54b1f1\") " Dec 02 11:08:47 crc kubenswrapper[4679]: I1202 11:08:47.999204 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-utilities" (OuterVolumeSpecName: "utilities") pod "3072c031-d396-444d-b7b6-b1085b54b1f1" (UID: "3072c031-d396-444d-b7b6-b1085b54b1f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.006532 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3072c031-d396-444d-b7b6-b1085b54b1f1-kube-api-access-5r86m" (OuterVolumeSpecName: "kube-api-access-5r86m") pod "3072c031-d396-444d-b7b6-b1085b54b1f1" (UID: "3072c031-d396-444d-b7b6-b1085b54b1f1"). InnerVolumeSpecName "kube-api-access-5r86m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.020051 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3072c031-d396-444d-b7b6-b1085b54b1f1" (UID: "3072c031-d396-444d-b7b6-b1085b54b1f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.100787 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.101393 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3072c031-d396-444d-b7b6-b1085b54b1f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.101430 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r86m\" (UniqueName: \"kubernetes.io/projected/3072c031-d396-444d-b7b6-b1085b54b1f1-kube-api-access-5r86m\") on node \"crc\" DevicePath \"\"" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.293965 4679 generic.go:334] "Generic (PLEG): container finished" podID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerID="e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc" exitCode=0 Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.294014 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v82j9" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.294012 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v82j9" event={"ID":"3072c031-d396-444d-b7b6-b1085b54b1f1","Type":"ContainerDied","Data":"e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc"} Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.294158 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v82j9" event={"ID":"3072c031-d396-444d-b7b6-b1085b54b1f1","Type":"ContainerDied","Data":"92a1edfee7342b2898f80c99839cd252fea48869c295a4253d82d646c807395e"} Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.294195 4679 scope.go:117] "RemoveContainer" containerID="e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.323555 4679 scope.go:117] "RemoveContainer" containerID="e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.330665 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v82j9"] Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.338619 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v82j9"] Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.355565 4679 scope.go:117] "RemoveContainer" containerID="0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.400482 4679 scope.go:117] "RemoveContainer" containerID="e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc" Dec 02 11:08:48 crc kubenswrapper[4679]: E1202 11:08:48.401413 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc\": container with ID starting with e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc not found: ID does not exist" containerID="e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.401447 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc"} err="failed to get container status \"e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc\": rpc error: code = NotFound desc = could not find container \"e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc\": container with ID starting with e50bb9bbb631d2db84ac977045c2514e1f324467cb9206a509cf258f002874bc not found: ID does not exist" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.401467 4679 scope.go:117] "RemoveContainer" containerID="e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29" Dec 02 11:08:48 crc kubenswrapper[4679]: E1202 11:08:48.401701 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29\": container with ID starting with e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29 not found: ID does not exist" containerID="e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.401718 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29"} err="failed to get container status \"e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29\": rpc error: code = NotFound desc = could not find container \"e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29\": container with ID starting with e0e1fe0894114228dccff00df66a8cbb71d7721bf8a556c185a1ac4e1e750a29 not found: ID does not exist" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.401735 4679 scope.go:117] "RemoveContainer" containerID="0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa" Dec 02 11:08:48 crc kubenswrapper[4679]: E1202 11:08:48.401982 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa\": container with ID starting with 0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa not found: ID does not exist" containerID="0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.401999 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa"} err="failed to get container status \"0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa\": rpc error: code = NotFound desc = could not find container \"0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa\": container with ID starting with 0ac0b88682c2162a001c7e07536ff6be99e80f2ff0ac3992ddae43522777f8fa not found: ID does not exist" Dec 02 11:08:48 crc kubenswrapper[4679]: I1202 11:08:48.923037 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" path="/var/lib/kubelet/pods/3072c031-d396-444d-b7b6-b1085b54b1f1/volumes" Dec 02 11:10:16 crc kubenswrapper[4679]: I1202 11:10:16.932083 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:10:16 crc kubenswrapper[4679]: I1202 11:10:16.932884 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:10:46 crc kubenswrapper[4679]: I1202 11:10:46.932488 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:10:46 crc kubenswrapper[4679]: I1202 11:10:46.933170 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:11:16 crc kubenswrapper[4679]: I1202 11:11:16.943997 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:11:16 crc kubenswrapper[4679]: I1202 11:11:16.944580 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:11:16 crc kubenswrapper[4679]: I1202 11:11:16.944647 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 11:11:16 crc kubenswrapper[4679]: I1202 11:11:16.945683 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 11:11:16 crc kubenswrapper[4679]: I1202 11:11:16.945784 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" gracePeriod=600 Dec 02 11:11:17 crc kubenswrapper[4679]: E1202 11:11:17.068128 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:11:17 crc kubenswrapper[4679]: I1202 11:11:17.751895 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" exitCode=0 Dec 02 11:11:17 crc kubenswrapper[4679]: I1202 11:11:17.751992 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a"} Dec 02 11:11:17 crc kubenswrapper[4679]: I1202 11:11:17.752436 4679 scope.go:117] "RemoveContainer" containerID="70127d6ce84a03c1c8b8d09d74c1b2753f1cfd4c549cf4f0d5c6e12b15b2f314" Dec 02 11:11:17 crc kubenswrapper[4679]: I1202 11:11:17.753427 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:11:17 crc kubenswrapper[4679]: E1202 11:11:17.754030 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:11:30 crc kubenswrapper[4679]: I1202 11:11:30.909514 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:11:30 crc kubenswrapper[4679]: E1202 11:11:30.910344 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:11:44 crc kubenswrapper[4679]: I1202 11:11:44.909140 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:11:44 crc kubenswrapper[4679]: E1202 11:11:44.909911 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:11:57 crc kubenswrapper[4679]: I1202 11:11:57.909453 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:11:57 crc kubenswrapper[4679]: E1202 11:11:57.910330 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:12:09 crc kubenswrapper[4679]: I1202 11:12:09.909432 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:12:09 crc kubenswrapper[4679]: E1202 11:12:09.910029 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:12:21 crc kubenswrapper[4679]: I1202 11:12:21.909641 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:12:21 crc kubenswrapper[4679]: E1202 11:12:21.910605 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:12:32 crc kubenswrapper[4679]: I1202 11:12:32.908501 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:12:32 crc kubenswrapper[4679]: E1202 11:12:32.909241 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:12:44 crc kubenswrapper[4679]: I1202 11:12:44.912541 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:12:44 crc kubenswrapper[4679]: E1202 11:12:44.913579 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:12:59 crc kubenswrapper[4679]: I1202 11:12:59.910013 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:12:59 crc kubenswrapper[4679]: E1202 11:12:59.910851 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:13:13 crc kubenswrapper[4679]: I1202 11:13:13.913236 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:13:13 crc kubenswrapper[4679]: E1202 11:13:13.915159 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:13:24 crc kubenswrapper[4679]: I1202 11:13:24.908502 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:13:24 crc kubenswrapper[4679]: E1202 11:13:24.909574 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:13:38 crc kubenswrapper[4679]: I1202 11:13:38.909584 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:13:38 crc kubenswrapper[4679]: E1202 11:13:38.910456 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:13:50 crc kubenswrapper[4679]: I1202 11:13:50.911275 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:13:50 crc kubenswrapper[4679]: E1202 11:13:50.912041 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:14:04 crc kubenswrapper[4679]: I1202 11:14:04.909139 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:14:04 crc kubenswrapper[4679]: E1202 11:14:04.911625 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:14:16 crc kubenswrapper[4679]: I1202 11:14:16.916884 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:14:16 crc kubenswrapper[4679]: E1202 11:14:16.917875 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:14:28 crc kubenswrapper[4679]: I1202 11:14:28.909288 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:14:28 crc kubenswrapper[4679]: E1202 11:14:28.910445 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.689439 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p58hn"] Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.689949 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="extract-utilities" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.689972 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="extract-utilities" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.689991 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="extract-utilities" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690000 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="extract-utilities" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690022 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690032 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690046 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="extract-content" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690054 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="extract-content" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690065 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="extract-content" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690072 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="extract-content" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690091 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="extract-content" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690099 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="extract-content" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690108 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690115 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690147 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="extract-utilities" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690155 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="extract-utilities" Dec 02 11:14:29 crc kubenswrapper[4679]: E1202 11:14:29.690171 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690181 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690397 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bfda4e3-b65e-40b4-93f8-c7891a983cce" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690411 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0938f55-510d-4ed1-b71a-05d48e6fd14c" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.690434 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3072c031-d396-444d-b7b6-b1085b54b1f1" containerName="registry-server" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.691946 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.751563 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p58hn"] Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.779430 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-utilities\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.779505 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-catalog-content\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.779530 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4898j\" (UniqueName: \"kubernetes.io/projected/e56417fc-7af1-452f-94dd-bdcfdca4878a-kube-api-access-4898j\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.884204 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-utilities\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.884256 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-catalog-content\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.884273 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4898j\" (UniqueName: \"kubernetes.io/projected/e56417fc-7af1-452f-94dd-bdcfdca4878a-kube-api-access-4898j\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.885055 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-utilities\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.885180 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-catalog-content\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:29 crc kubenswrapper[4679]: I1202 11:14:29.905934 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4898j\" (UniqueName: \"kubernetes.io/projected/e56417fc-7af1-452f-94dd-bdcfdca4878a-kube-api-access-4898j\") pod \"redhat-operators-p58hn\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:30 crc kubenswrapper[4679]: I1202 11:14:30.026325 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:30 crc kubenswrapper[4679]: I1202 11:14:30.562997 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p58hn"] Dec 02 11:14:30 crc kubenswrapper[4679]: I1202 11:14:30.679341 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerStarted","Data":"65f0ad1e9f015e69ec17409c3e982965e6a921627f0745bf514fd2d57bc02b15"} Dec 02 11:14:31 crc kubenswrapper[4679]: I1202 11:14:31.701504 4679 generic.go:334] "Generic (PLEG): container finished" podID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerID="029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339" exitCode=0 Dec 02 11:14:31 crc kubenswrapper[4679]: I1202 11:14:31.701597 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerDied","Data":"029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339"} Dec 02 11:14:31 crc kubenswrapper[4679]: I1202 11:14:31.705654 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 11:14:32 crc kubenswrapper[4679]: I1202 11:14:32.713968 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerStarted","Data":"ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554"} Dec 02 11:14:33 crc kubenswrapper[4679]: I1202 11:14:33.725785 4679 generic.go:334] "Generic (PLEG): container finished" podID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerID="ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554" exitCode=0 Dec 02 11:14:33 crc kubenswrapper[4679]: I1202 11:14:33.726108 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerDied","Data":"ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554"} Dec 02 11:14:36 crc kubenswrapper[4679]: I1202 11:14:36.761640 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerStarted","Data":"6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e"} Dec 02 11:14:36 crc kubenswrapper[4679]: I1202 11:14:36.790478 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p58hn" podStartSLOduration=4.5384536109999996 podStartE2EDuration="7.790452192s" podCreationTimestamp="2025-12-02 11:14:29 +0000 UTC" firstStartedPulling="2025-12-02 11:14:31.705376988 +0000 UTC m=+3325.035515858" lastFinishedPulling="2025-12-02 11:14:34.957375579 +0000 UTC m=+3328.287514439" observedRunningTime="2025-12-02 11:14:36.786585442 +0000 UTC m=+3330.116724352" watchObservedRunningTime="2025-12-02 11:14:36.790452192 +0000 UTC m=+3330.120591062" Dec 02 11:14:40 crc kubenswrapper[4679]: I1202 11:14:40.027413 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:40 crc kubenswrapper[4679]: I1202 11:14:40.027746 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:41 crc kubenswrapper[4679]: I1202 11:14:41.072562 4679 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p58hn" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="registry-server" probeResult="failure" output=< Dec 02 11:14:41 crc kubenswrapper[4679]: timeout: failed to connect service ":50051" within 1s Dec 02 11:14:41 crc kubenswrapper[4679]: > Dec 02 11:14:42 crc kubenswrapper[4679]: I1202 11:14:42.908777 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:14:42 crc kubenswrapper[4679]: E1202 11:14:42.910399 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:14:50 crc kubenswrapper[4679]: I1202 11:14:50.088415 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:50 crc kubenswrapper[4679]: I1202 11:14:50.171553 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:52 crc kubenswrapper[4679]: I1202 11:14:52.930826 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p58hn"] Dec 02 11:14:52 crc kubenswrapper[4679]: I1202 11:14:52.931790 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p58hn" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="registry-server" containerID="cri-o://6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e" gracePeriod=2 Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.454781 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.506100 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-catalog-content\") pod \"e56417fc-7af1-452f-94dd-bdcfdca4878a\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.506230 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4898j\" (UniqueName: \"kubernetes.io/projected/e56417fc-7af1-452f-94dd-bdcfdca4878a-kube-api-access-4898j\") pod \"e56417fc-7af1-452f-94dd-bdcfdca4878a\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.506329 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-utilities\") pod \"e56417fc-7af1-452f-94dd-bdcfdca4878a\" (UID: \"e56417fc-7af1-452f-94dd-bdcfdca4878a\") " Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.507646 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-utilities" (OuterVolumeSpecName: "utilities") pod "e56417fc-7af1-452f-94dd-bdcfdca4878a" (UID: "e56417fc-7af1-452f-94dd-bdcfdca4878a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.516659 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e56417fc-7af1-452f-94dd-bdcfdca4878a-kube-api-access-4898j" (OuterVolumeSpecName: "kube-api-access-4898j") pod "e56417fc-7af1-452f-94dd-bdcfdca4878a" (UID: "e56417fc-7af1-452f-94dd-bdcfdca4878a"). InnerVolumeSpecName "kube-api-access-4898j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.609366 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4898j\" (UniqueName: \"kubernetes.io/projected/e56417fc-7af1-452f-94dd-bdcfdca4878a-kube-api-access-4898j\") on node \"crc\" DevicePath \"\"" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.609411 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.613256 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e56417fc-7af1-452f-94dd-bdcfdca4878a" (UID: "e56417fc-7af1-452f-94dd-bdcfdca4878a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.711251 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e56417fc-7af1-452f-94dd-bdcfdca4878a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.930424 4679 generic.go:334] "Generic (PLEG): container finished" podID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerID="6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e" exitCode=0 Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.930481 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerDied","Data":"6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e"} Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.930542 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p58hn" event={"ID":"e56417fc-7af1-452f-94dd-bdcfdca4878a","Type":"ContainerDied","Data":"65f0ad1e9f015e69ec17409c3e982965e6a921627f0745bf514fd2d57bc02b15"} Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.930564 4679 scope.go:117] "RemoveContainer" containerID="6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.930566 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p58hn" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.954714 4679 scope.go:117] "RemoveContainer" containerID="ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554" Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.978654 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p58hn"] Dec 02 11:14:53 crc kubenswrapper[4679]: I1202 11:14:53.991071 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p58hn"] Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.008050 4679 scope.go:117] "RemoveContainer" containerID="029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.047082 4679 scope.go:117] "RemoveContainer" containerID="6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e" Dec 02 11:14:54 crc kubenswrapper[4679]: E1202 11:14:54.047524 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e\": container with ID starting with 6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e not found: ID does not exist" containerID="6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.047584 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e"} err="failed to get container status \"6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e\": rpc error: code = NotFound desc = could not find container \"6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e\": container with ID starting with 6fdfe1ab3d92c77f53dd985abaec8db15bfa9e39dd8739a05766bfe7beaf566e not found: ID does not exist" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.047685 4679 scope.go:117] "RemoveContainer" containerID="ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554" Dec 02 11:14:54 crc kubenswrapper[4679]: E1202 11:14:54.047966 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554\": container with ID starting with ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554 not found: ID does not exist" containerID="ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.048002 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554"} err="failed to get container status \"ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554\": rpc error: code = NotFound desc = could not find container \"ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554\": container with ID starting with ee28cc4dd5cc2db2edbcdf8e864a48e8c7a4e7931c3429f62f16ff313ae93554 not found: ID does not exist" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.048018 4679 scope.go:117] "RemoveContainer" containerID="029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339" Dec 02 11:14:54 crc kubenswrapper[4679]: E1202 11:14:54.048277 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339\": container with ID starting with 029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339 not found: ID does not exist" containerID="029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.048315 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339"} err="failed to get container status \"029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339\": rpc error: code = NotFound desc = could not find container \"029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339\": container with ID starting with 029254fd29e37034a7a928190202c508fbb5e641405e7ed1cf315b751b62e339 not found: ID does not exist" Dec 02 11:14:54 crc kubenswrapper[4679]: I1202 11:14:54.925323 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" path="/var/lib/kubelet/pods/e56417fc-7af1-452f-94dd-bdcfdca4878a/volumes" Dec 02 11:14:55 crc kubenswrapper[4679]: I1202 11:14:55.909507 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:14:55 crc kubenswrapper[4679]: E1202 11:14:55.909752 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.159217 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk"] Dec 02 11:15:00 crc kubenswrapper[4679]: E1202 11:15:00.160429 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="extract-content" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.160449 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="extract-content" Dec 02 11:15:00 crc kubenswrapper[4679]: E1202 11:15:00.160464 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="registry-server" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.160472 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="registry-server" Dec 02 11:15:00 crc kubenswrapper[4679]: E1202 11:15:00.160511 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="extract-utilities" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.160520 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="extract-utilities" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.160774 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56417fc-7af1-452f-94dd-bdcfdca4878a" containerName="registry-server" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.161628 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.164600 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.164844 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.176335 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk"] Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.243036 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e9bb0fd-3b09-4303-8cef-23312ef2316d-secret-volume\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.243102 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e9bb0fd-3b09-4303-8cef-23312ef2316d-config-volume\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.243349 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lp6t\" (UniqueName: \"kubernetes.io/projected/3e9bb0fd-3b09-4303-8cef-23312ef2316d-kube-api-access-6lp6t\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.345475 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e9bb0fd-3b09-4303-8cef-23312ef2316d-secret-volume\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.345541 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e9bb0fd-3b09-4303-8cef-23312ef2316d-config-volume\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.345608 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lp6t\" (UniqueName: \"kubernetes.io/projected/3e9bb0fd-3b09-4303-8cef-23312ef2316d-kube-api-access-6lp6t\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.346822 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e9bb0fd-3b09-4303-8cef-23312ef2316d-config-volume\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.354346 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e9bb0fd-3b09-4303-8cef-23312ef2316d-secret-volume\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.368588 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lp6t\" (UniqueName: \"kubernetes.io/projected/3e9bb0fd-3b09-4303-8cef-23312ef2316d-kube-api-access-6lp6t\") pod \"collect-profiles-29411235-xlsxk\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.478939 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.937668 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk"] Dec 02 11:15:00 crc kubenswrapper[4679]: I1202 11:15:00.993433 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" event={"ID":"3e9bb0fd-3b09-4303-8cef-23312ef2316d","Type":"ContainerStarted","Data":"9c268ed9250cddd3a43b2c88ed2d6e78afdb0d4269ea2bf2fa8174ee93aaae0f"} Dec 02 11:15:02 crc kubenswrapper[4679]: I1202 11:15:02.004541 4679 generic.go:334] "Generic (PLEG): container finished" podID="3e9bb0fd-3b09-4303-8cef-23312ef2316d" containerID="bcfabb295f980aab422d0dbc34428ae15fd3105d2e4dea9404f0b60c2262dba0" exitCode=0 Dec 02 11:15:02 crc kubenswrapper[4679]: I1202 11:15:02.004612 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" event={"ID":"3e9bb0fd-3b09-4303-8cef-23312ef2316d","Type":"ContainerDied","Data":"bcfabb295f980aab422d0dbc34428ae15fd3105d2e4dea9404f0b60c2262dba0"} Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.374392 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.410335 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e9bb0fd-3b09-4303-8cef-23312ef2316d-secret-volume\") pod \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.410556 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e9bb0fd-3b09-4303-8cef-23312ef2316d-config-volume\") pod \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.410585 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lp6t\" (UniqueName: \"kubernetes.io/projected/3e9bb0fd-3b09-4303-8cef-23312ef2316d-kube-api-access-6lp6t\") pod \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\" (UID: \"3e9bb0fd-3b09-4303-8cef-23312ef2316d\") " Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.412018 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e9bb0fd-3b09-4303-8cef-23312ef2316d-config-volume" (OuterVolumeSpecName: "config-volume") pod "3e9bb0fd-3b09-4303-8cef-23312ef2316d" (UID: "3e9bb0fd-3b09-4303-8cef-23312ef2316d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.419875 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e9bb0fd-3b09-4303-8cef-23312ef2316d-kube-api-access-6lp6t" (OuterVolumeSpecName: "kube-api-access-6lp6t") pod "3e9bb0fd-3b09-4303-8cef-23312ef2316d" (UID: "3e9bb0fd-3b09-4303-8cef-23312ef2316d"). InnerVolumeSpecName "kube-api-access-6lp6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.420588 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e9bb0fd-3b09-4303-8cef-23312ef2316d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3e9bb0fd-3b09-4303-8cef-23312ef2316d" (UID: "3e9bb0fd-3b09-4303-8cef-23312ef2316d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.513431 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e9bb0fd-3b09-4303-8cef-23312ef2316d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.513476 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lp6t\" (UniqueName: \"kubernetes.io/projected/3e9bb0fd-3b09-4303-8cef-23312ef2316d-kube-api-access-6lp6t\") on node \"crc\" DevicePath \"\"" Dec 02 11:15:03 crc kubenswrapper[4679]: I1202 11:15:03.513491 4679 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e9bb0fd-3b09-4303-8cef-23312ef2316d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 11:15:04 crc kubenswrapper[4679]: I1202 11:15:04.024747 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" event={"ID":"3e9bb0fd-3b09-4303-8cef-23312ef2316d","Type":"ContainerDied","Data":"9c268ed9250cddd3a43b2c88ed2d6e78afdb0d4269ea2bf2fa8174ee93aaae0f"} Dec 02 11:15:04 crc kubenswrapper[4679]: I1202 11:15:04.024795 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411235-xlsxk" Dec 02 11:15:04 crc kubenswrapper[4679]: I1202 11:15:04.024807 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c268ed9250cddd3a43b2c88ed2d6e78afdb0d4269ea2bf2fa8174ee93aaae0f" Dec 02 11:15:04 crc kubenswrapper[4679]: I1202 11:15:04.447511 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n"] Dec 02 11:15:04 crc kubenswrapper[4679]: I1202 11:15:04.457334 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411190-l245n"] Dec 02 11:15:04 crc kubenswrapper[4679]: I1202 11:15:04.925669 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18bd6ce5-4362-4262-99f9-cef7c2f0d104" path="/var/lib/kubelet/pods/18bd6ce5-4362-4262-99f9-cef7c2f0d104/volumes" Dec 02 11:15:07 crc kubenswrapper[4679]: I1202 11:15:07.908993 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:15:07 crc kubenswrapper[4679]: E1202 11:15:07.909750 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:15:18 crc kubenswrapper[4679]: I1202 11:15:18.908709 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:15:18 crc kubenswrapper[4679]: E1202 11:15:18.909605 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:15:23 crc kubenswrapper[4679]: I1202 11:15:23.387342 4679 scope.go:117] "RemoveContainer" containerID="b4a04b64c9e2ac8bd7aa4240181851555fdb3925ca820823fe1d57a6e9de3945" Dec 02 11:15:30 crc kubenswrapper[4679]: I1202 11:15:30.909400 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:15:30 crc kubenswrapper[4679]: E1202 11:15:30.910707 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:15:44 crc kubenswrapper[4679]: I1202 11:15:44.929165 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:15:44 crc kubenswrapper[4679]: E1202 11:15:44.930796 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:15:55 crc kubenswrapper[4679]: I1202 11:15:55.910536 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:15:55 crc kubenswrapper[4679]: E1202 11:15:55.911438 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:16:10 crc kubenswrapper[4679]: I1202 11:16:10.908930 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:16:10 crc kubenswrapper[4679]: E1202 11:16:10.909719 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:16:23 crc kubenswrapper[4679]: I1202 11:16:23.909145 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:16:24 crc kubenswrapper[4679]: I1202 11:16:24.833504 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"68ca6b38ae7ce6a11d6cd6e7bfb27a17db2519f3b4ecdb2f0b95937e41fad9d0"} Dec 02 11:16:54 crc kubenswrapper[4679]: I1202 11:16:54.114600 4679 generic.go:334] "Generic (PLEG): container finished" podID="564a5130-acc1-40e0-814b-2e9bfd2d967d" containerID="8914a691fce4dac0dd056eef6fe9098a82781955baca29357adfc15601292d2c" exitCode=0 Dec 02 11:16:54 crc kubenswrapper[4679]: I1202 11:16:54.114742 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564a5130-acc1-40e0-814b-2e9bfd2d967d","Type":"ContainerDied","Data":"8914a691fce4dac0dd056eef6fe9098a82781955baca29357adfc15601292d2c"} Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.545990 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.716716 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.716775 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ca-certs\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.716855 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vslz\" (UniqueName: \"kubernetes.io/projected/564a5130-acc1-40e0-814b-2e9bfd2d967d-kube-api-access-9vslz\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.716916 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.716982 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config-secret\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.717041 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-workdir\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.717072 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-temporary\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.717184 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-config-data\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.717209 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ssh-key\") pod \"564a5130-acc1-40e0-814b-2e9bfd2d967d\" (UID: \"564a5130-acc1-40e0-814b-2e9bfd2d967d\") " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.718007 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.718418 4679 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.718473 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-config-data" (OuterVolumeSpecName: "config-data") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.722944 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.725285 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/564a5130-acc1-40e0-814b-2e9bfd2d967d-kube-api-access-9vslz" (OuterVolumeSpecName: "kube-api-access-9vslz") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "kube-api-access-9vslz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.725328 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.760516 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.768071 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.780345 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.786818 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "564a5130-acc1-40e0-814b-2e9bfd2d967d" (UID: "564a5130-acc1-40e0-814b-2e9bfd2d967d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821546 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vslz\" (UniqueName: \"kubernetes.io/projected/564a5130-acc1-40e0-814b-2e9bfd2d967d-kube-api-access-9vslz\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821579 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821591 4679 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821604 4679 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564a5130-acc1-40e0-814b-2e9bfd2d967d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821617 4679 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564a5130-acc1-40e0-814b-2e9bfd2d967d-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821630 4679 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821669 4679 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.821680 4679 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564a5130-acc1-40e0-814b-2e9bfd2d967d-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.844509 4679 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 02 11:16:55 crc kubenswrapper[4679]: I1202 11:16:55.923695 4679 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 02 11:16:56 crc kubenswrapper[4679]: I1202 11:16:56.141763 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564a5130-acc1-40e0-814b-2e9bfd2d967d","Type":"ContainerDied","Data":"39f51430602728ed632223b28025488d5a4e060594fd9447b0b4a890349f37a3"} Dec 02 11:16:56 crc kubenswrapper[4679]: I1202 11:16:56.142139 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39f51430602728ed632223b28025488d5a4e060594fd9447b0b4a890349f37a3" Dec 02 11:16:56 crc kubenswrapper[4679]: I1202 11:16:56.141821 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.729399 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 11:17:07 crc kubenswrapper[4679]: E1202 11:17:07.730432 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e9bb0fd-3b09-4303-8cef-23312ef2316d" containerName="collect-profiles" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.730451 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e9bb0fd-3b09-4303-8cef-23312ef2316d" containerName="collect-profiles" Dec 02 11:17:07 crc kubenswrapper[4679]: E1202 11:17:07.730500 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564a5130-acc1-40e0-814b-2e9bfd2d967d" containerName="tempest-tests-tempest-tests-runner" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.730507 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="564a5130-acc1-40e0-814b-2e9bfd2d967d" containerName="tempest-tests-tempest-tests-runner" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.730688 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e9bb0fd-3b09-4303-8cef-23312ef2316d" containerName="collect-profiles" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.730712 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="564a5130-acc1-40e0-814b-2e9bfd2d967d" containerName="tempest-tests-tempest-tests-runner" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.731277 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.733791 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5xprf" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.739381 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.876738 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.876819 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-625g6\" (UniqueName: \"kubernetes.io/projected/d17cc270-8f47-4ea2-ba89-6bb3011ed620-kube-api-access-625g6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.979645 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.980354 4679 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:07 crc kubenswrapper[4679]: I1202 11:17:07.980553 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-625g6\" (UniqueName: \"kubernetes.io/projected/d17cc270-8f47-4ea2-ba89-6bb3011ed620-kube-api-access-625g6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:08 crc kubenswrapper[4679]: I1202 11:17:08.016032 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:08 crc kubenswrapper[4679]: I1202 11:17:08.016931 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-625g6\" (UniqueName: \"kubernetes.io/projected/d17cc270-8f47-4ea2-ba89-6bb3011ed620-kube-api-access-625g6\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d17cc270-8f47-4ea2-ba89-6bb3011ed620\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:08 crc kubenswrapper[4679]: I1202 11:17:08.054956 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 11:17:08 crc kubenswrapper[4679]: I1202 11:17:08.514150 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 11:17:09 crc kubenswrapper[4679]: I1202 11:17:09.305864 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d17cc270-8f47-4ea2-ba89-6bb3011ed620","Type":"ContainerStarted","Data":"614646365bf17b7ccef7d7a1d3f22e30e3bf5a41ece58c1cccfd7dfdb06eaf1c"} Dec 02 11:17:10 crc kubenswrapper[4679]: I1202 11:17:10.316952 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d17cc270-8f47-4ea2-ba89-6bb3011ed620","Type":"ContainerStarted","Data":"71f9a194313ce4ed0e7c615f977abffb21e856e95126dcd82e17c6e1d009da1c"} Dec 02 11:17:10 crc kubenswrapper[4679]: I1202 11:17:10.334209 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.417285569 podStartE2EDuration="3.334194106s" podCreationTimestamp="2025-12-02 11:17:07 +0000 UTC" firstStartedPulling="2025-12-02 11:17:08.514719751 +0000 UTC m=+3481.844858611" lastFinishedPulling="2025-12-02 11:17:09.431628288 +0000 UTC m=+3482.761767148" observedRunningTime="2025-12-02 11:17:10.332260141 +0000 UTC m=+3483.662399001" watchObservedRunningTime="2025-12-02 11:17:10.334194106 +0000 UTC m=+3483.664332966" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.272187 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-529wn/must-gather-x86c5"] Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.274331 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.279884 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-529wn"/"openshift-service-ca.crt" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.282282 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-529wn"/"kube-root-ca.crt" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.302657 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-529wn/must-gather-x86c5"] Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.391563 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxglw\" (UniqueName: \"kubernetes.io/projected/2d48b322-087a-4173-8110-262905e0d358-kube-api-access-lxglw\") pod \"must-gather-x86c5\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.391611 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d48b322-087a-4173-8110-262905e0d358-must-gather-output\") pod \"must-gather-x86c5\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.493245 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxglw\" (UniqueName: \"kubernetes.io/projected/2d48b322-087a-4173-8110-262905e0d358-kube-api-access-lxglw\") pod \"must-gather-x86c5\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.493579 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d48b322-087a-4173-8110-262905e0d358-must-gather-output\") pod \"must-gather-x86c5\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.493903 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d48b322-087a-4173-8110-262905e0d358-must-gather-output\") pod \"must-gather-x86c5\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.510879 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxglw\" (UniqueName: \"kubernetes.io/projected/2d48b322-087a-4173-8110-262905e0d358-kube-api-access-lxglw\") pod \"must-gather-x86c5\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:32 crc kubenswrapper[4679]: I1202 11:17:32.602747 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:17:33 crc kubenswrapper[4679]: I1202 11:17:33.096325 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-529wn/must-gather-x86c5"] Dec 02 11:17:33 crc kubenswrapper[4679]: I1202 11:17:33.542479 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/must-gather-x86c5" event={"ID":"2d48b322-087a-4173-8110-262905e0d358","Type":"ContainerStarted","Data":"f65f7e6edd7d5dc924a77f6aee9449c263238bf989820fabecacd69b2a45a1c5"} Dec 02 11:17:37 crc kubenswrapper[4679]: I1202 11:17:37.584204 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/must-gather-x86c5" event={"ID":"2d48b322-087a-4173-8110-262905e0d358","Type":"ContainerStarted","Data":"796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f"} Dec 02 11:17:37 crc kubenswrapper[4679]: I1202 11:17:37.585498 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/must-gather-x86c5" event={"ID":"2d48b322-087a-4173-8110-262905e0d358","Type":"ContainerStarted","Data":"5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0"} Dec 02 11:17:37 crc kubenswrapper[4679]: I1202 11:17:37.612782 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-529wn/must-gather-x86c5" podStartSLOduration=2.065531623 podStartE2EDuration="5.61275561s" podCreationTimestamp="2025-12-02 11:17:32 +0000 UTC" firstStartedPulling="2025-12-02 11:17:33.107626187 +0000 UTC m=+3506.437765047" lastFinishedPulling="2025-12-02 11:17:36.654850164 +0000 UTC m=+3509.984989034" observedRunningTime="2025-12-02 11:17:37.598406291 +0000 UTC m=+3510.928545181" watchObservedRunningTime="2025-12-02 11:17:37.61275561 +0000 UTC m=+3510.942894510" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.271867 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-529wn/crc-debug-fc4mh"] Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.273534 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.276071 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-529wn"/"default-dockercfg-qrd6q" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.455618 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-host\") pod \"crc-debug-fc4mh\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.456185 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdjwj\" (UniqueName: \"kubernetes.io/projected/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-kube-api-access-gdjwj\") pod \"crc-debug-fc4mh\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.558148 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-host\") pod \"crc-debug-fc4mh\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.558255 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdjwj\" (UniqueName: \"kubernetes.io/projected/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-kube-api-access-gdjwj\") pod \"crc-debug-fc4mh\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.558299 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-host\") pod \"crc-debug-fc4mh\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.578887 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdjwj\" (UniqueName: \"kubernetes.io/projected/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-kube-api-access-gdjwj\") pod \"crc-debug-fc4mh\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: I1202 11:17:40.590335 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:17:40 crc kubenswrapper[4679]: W1202 11:17:40.619094 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod512cf436_5b7f_4d5c_aa4e_6ca00ca0b474.slice/crio-3e911dfe6ad793e8c0f0d908deeafcbfdaf6bf29d182431a3f2dff6a2102de69 WatchSource:0}: Error finding container 3e911dfe6ad793e8c0f0d908deeafcbfdaf6bf29d182431a3f2dff6a2102de69: Status 404 returned error can't find the container with id 3e911dfe6ad793e8c0f0d908deeafcbfdaf6bf29d182431a3f2dff6a2102de69 Dec 02 11:17:41 crc kubenswrapper[4679]: I1202 11:17:41.621504 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-fc4mh" event={"ID":"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474","Type":"ContainerStarted","Data":"3e911dfe6ad793e8c0f0d908deeafcbfdaf6bf29d182431a3f2dff6a2102de69"} Dec 02 11:17:51 crc kubenswrapper[4679]: I1202 11:17:51.716868 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-fc4mh" event={"ID":"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474","Type":"ContainerStarted","Data":"47958b45aabc226c8d4431a181963d1eda9445c5a7d92444d455858369c86c35"} Dec 02 11:17:51 crc kubenswrapper[4679]: I1202 11:17:51.736537 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-529wn/crc-debug-fc4mh" podStartSLOduration=1.66803105 podStartE2EDuration="11.736517921s" podCreationTimestamp="2025-12-02 11:17:40 +0000 UTC" firstStartedPulling="2025-12-02 11:17:40.621379624 +0000 UTC m=+3513.951518484" lastFinishedPulling="2025-12-02 11:17:50.689866495 +0000 UTC m=+3524.020005355" observedRunningTime="2025-12-02 11:17:51.730904051 +0000 UTC m=+3525.061042901" watchObservedRunningTime="2025-12-02 11:17:51.736517921 +0000 UTC m=+3525.066656781" Dec 02 11:18:30 crc kubenswrapper[4679]: I1202 11:18:30.113127 4679 generic.go:334] "Generic (PLEG): container finished" podID="512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" containerID="47958b45aabc226c8d4431a181963d1eda9445c5a7d92444d455858369c86c35" exitCode=0 Dec 02 11:18:30 crc kubenswrapper[4679]: I1202 11:18:30.113240 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-fc4mh" event={"ID":"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474","Type":"ContainerDied","Data":"47958b45aabc226c8d4431a181963d1eda9445c5a7d92444d455858369c86c35"} Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.223064 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.261179 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-529wn/crc-debug-fc4mh"] Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.269051 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-529wn/crc-debug-fc4mh"] Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.331737 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-host\") pod \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.331802 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdjwj\" (UniqueName: \"kubernetes.io/projected/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-kube-api-access-gdjwj\") pod \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\" (UID: \"512cf436-5b7f-4d5c-aa4e-6ca00ca0b474\") " Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.332078 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-host" (OuterVolumeSpecName: "host") pod "512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" (UID: "512cf436-5b7f-4d5c-aa4e-6ca00ca0b474"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.332365 4679 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-host\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.337946 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-kube-api-access-gdjwj" (OuterVolumeSpecName: "kube-api-access-gdjwj") pod "512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" (UID: "512cf436-5b7f-4d5c-aa4e-6ca00ca0b474"). InnerVolumeSpecName "kube-api-access-gdjwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:18:31 crc kubenswrapper[4679]: I1202 11:18:31.434367 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdjwj\" (UniqueName: \"kubernetes.io/projected/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474-kube-api-access-gdjwj\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.152065 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e911dfe6ad793e8c0f0d908deeafcbfdaf6bf29d182431a3f2dff6a2102de69" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.152217 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-fc4mh" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.444887 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-529wn/crc-debug-gp9jx"] Dec 02 11:18:32 crc kubenswrapper[4679]: E1202 11:18:32.445364 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" containerName="container-00" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.445378 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" containerName="container-00" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.445602 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" containerName="container-00" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.446346 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.448797 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-529wn"/"default-dockercfg-qrd6q" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.558263 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tw5v\" (UniqueName: \"kubernetes.io/projected/b1945e1d-60eb-471e-aedf-941e32244556-kube-api-access-5tw5v\") pod \"crc-debug-gp9jx\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.558462 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1945e1d-60eb-471e-aedf-941e32244556-host\") pod \"crc-debug-gp9jx\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.660698 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1945e1d-60eb-471e-aedf-941e32244556-host\") pod \"crc-debug-gp9jx\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.660734 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1945e1d-60eb-471e-aedf-941e32244556-host\") pod \"crc-debug-gp9jx\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.661141 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tw5v\" (UniqueName: \"kubernetes.io/projected/b1945e1d-60eb-471e-aedf-941e32244556-kube-api-access-5tw5v\") pod \"crc-debug-gp9jx\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.678263 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tw5v\" (UniqueName: \"kubernetes.io/projected/b1945e1d-60eb-471e-aedf-941e32244556-kube-api-access-5tw5v\") pod \"crc-debug-gp9jx\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.770323 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:32 crc kubenswrapper[4679]: I1202 11:18:32.920018 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="512cf436-5b7f-4d5c-aa4e-6ca00ca0b474" path="/var/lib/kubelet/pods/512cf436-5b7f-4d5c-aa4e-6ca00ca0b474/volumes" Dec 02 11:18:33 crc kubenswrapper[4679]: I1202 11:18:33.168480 4679 generic.go:334] "Generic (PLEG): container finished" podID="b1945e1d-60eb-471e-aedf-941e32244556" containerID="11987d5384a76c71bf60ff4829bf22874d7b95c07a29a4f12e8b1336301f1402" exitCode=0 Dec 02 11:18:33 crc kubenswrapper[4679]: I1202 11:18:33.168582 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-gp9jx" event={"ID":"b1945e1d-60eb-471e-aedf-941e32244556","Type":"ContainerDied","Data":"11987d5384a76c71bf60ff4829bf22874d7b95c07a29a4f12e8b1336301f1402"} Dec 02 11:18:33 crc kubenswrapper[4679]: I1202 11:18:33.169021 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-gp9jx" event={"ID":"b1945e1d-60eb-471e-aedf-941e32244556","Type":"ContainerStarted","Data":"2794f935ba3d00525d07b78e489c95a4c2f7d013f8d4e973e6144c4cd69414c1"} Dec 02 11:18:33 crc kubenswrapper[4679]: I1202 11:18:33.710569 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-529wn/crc-debug-gp9jx"] Dec 02 11:18:33 crc kubenswrapper[4679]: I1202 11:18:33.720954 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-529wn/crc-debug-gp9jx"] Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.284039 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.389628 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1945e1d-60eb-471e-aedf-941e32244556-host\") pod \"b1945e1d-60eb-471e-aedf-941e32244556\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.389720 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tw5v\" (UniqueName: \"kubernetes.io/projected/b1945e1d-60eb-471e-aedf-941e32244556-kube-api-access-5tw5v\") pod \"b1945e1d-60eb-471e-aedf-941e32244556\" (UID: \"b1945e1d-60eb-471e-aedf-941e32244556\") " Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.389784 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b1945e1d-60eb-471e-aedf-941e32244556-host" (OuterVolumeSpecName: "host") pod "b1945e1d-60eb-471e-aedf-941e32244556" (UID: "b1945e1d-60eb-471e-aedf-941e32244556"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.390684 4679 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b1945e1d-60eb-471e-aedf-941e32244556-host\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.398800 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1945e1d-60eb-471e-aedf-941e32244556-kube-api-access-5tw5v" (OuterVolumeSpecName: "kube-api-access-5tw5v") pod "b1945e1d-60eb-471e-aedf-941e32244556" (UID: "b1945e1d-60eb-471e-aedf-941e32244556"). InnerVolumeSpecName "kube-api-access-5tw5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.492960 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tw5v\" (UniqueName: \"kubernetes.io/projected/b1945e1d-60eb-471e-aedf-941e32244556-kube-api-access-5tw5v\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.919165 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1945e1d-60eb-471e-aedf-941e32244556" path="/var/lib/kubelet/pods/b1945e1d-60eb-471e-aedf-941e32244556/volumes" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.919770 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-529wn/crc-debug-c79gt"] Dec 02 11:18:34 crc kubenswrapper[4679]: E1202 11:18:34.920049 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1945e1d-60eb-471e-aedf-941e32244556" containerName="container-00" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.920060 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1945e1d-60eb-471e-aedf-941e32244556" containerName="container-00" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.920252 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1945e1d-60eb-471e-aedf-941e32244556" containerName="container-00" Dec 02 11:18:34 crc kubenswrapper[4679]: I1202 11:18:34.920847 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.001880 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d22a9d-a20f-460d-b030-63792ec5da2c-host\") pod \"crc-debug-c79gt\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.001940 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7sds\" (UniqueName: \"kubernetes.io/projected/36d22a9d-a20f-460d-b030-63792ec5da2c-kube-api-access-v7sds\") pod \"crc-debug-c79gt\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.105364 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d22a9d-a20f-460d-b030-63792ec5da2c-host\") pod \"crc-debug-c79gt\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.105424 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7sds\" (UniqueName: \"kubernetes.io/projected/36d22a9d-a20f-460d-b030-63792ec5da2c-kube-api-access-v7sds\") pod \"crc-debug-c79gt\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.105571 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d22a9d-a20f-460d-b030-63792ec5da2c-host\") pod \"crc-debug-c79gt\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.123532 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7sds\" (UniqueName: \"kubernetes.io/projected/36d22a9d-a20f-460d-b030-63792ec5da2c-kube-api-access-v7sds\") pod \"crc-debug-c79gt\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.190491 4679 scope.go:117] "RemoveContainer" containerID="11987d5384a76c71bf60ff4829bf22874d7b95c07a29a4f12e8b1336301f1402" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.190519 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-gp9jx" Dec 02 11:18:35 crc kubenswrapper[4679]: I1202 11:18:35.241821 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:35 crc kubenswrapper[4679]: W1202 11:18:35.271926 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36d22a9d_a20f_460d_b030_63792ec5da2c.slice/crio-dbb78ebd6ee6eb88f79c381a25faee3dba8322d43524269603eea2f1c3cfe351 WatchSource:0}: Error finding container dbb78ebd6ee6eb88f79c381a25faee3dba8322d43524269603eea2f1c3cfe351: Status 404 returned error can't find the container with id dbb78ebd6ee6eb88f79c381a25faee3dba8322d43524269603eea2f1c3cfe351 Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.201767 4679 generic.go:334] "Generic (PLEG): container finished" podID="36d22a9d-a20f-460d-b030-63792ec5da2c" containerID="ac5e6552023906620c4859520b8e6f014bd832302d9afadb24e5f14dc3e67370" exitCode=0 Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.201879 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-c79gt" event={"ID":"36d22a9d-a20f-460d-b030-63792ec5da2c","Type":"ContainerDied","Data":"ac5e6552023906620c4859520b8e6f014bd832302d9afadb24e5f14dc3e67370"} Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.202091 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/crc-debug-c79gt" event={"ID":"36d22a9d-a20f-460d-b030-63792ec5da2c","Type":"ContainerStarted","Data":"dbb78ebd6ee6eb88f79c381a25faee3dba8322d43524269603eea2f1c3cfe351"} Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.245885 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-529wn/crc-debug-c79gt"] Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.256855 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-529wn/crc-debug-c79gt"] Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.887409 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f896p"] Dec 02 11:18:36 crc kubenswrapper[4679]: E1202 11:18:36.888265 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d22a9d-a20f-460d-b030-63792ec5da2c" containerName="container-00" Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.888280 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d22a9d-a20f-460d-b030-63792ec5da2c" containerName="container-00" Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.888544 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d22a9d-a20f-460d-b030-63792ec5da2c" containerName="container-00" Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.890391 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.893565 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f896p"] Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.941387 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-catalog-content\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.941472 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjmwb\" (UniqueName: \"kubernetes.io/projected/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-kube-api-access-vjmwb\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:36 crc kubenswrapper[4679]: I1202 11:18:36.941500 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-utilities\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.043480 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-catalog-content\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.043589 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjmwb\" (UniqueName: \"kubernetes.io/projected/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-kube-api-access-vjmwb\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.043618 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-utilities\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.044425 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-utilities\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.044686 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-catalog-content\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.064293 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjmwb\" (UniqueName: \"kubernetes.io/projected/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-kube-api-access-vjmwb\") pod \"certified-operators-f896p\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.228213 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.312935 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.450867 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d22a9d-a20f-460d-b030-63792ec5da2c-host\") pod \"36d22a9d-a20f-460d-b030-63792ec5da2c\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.450969 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7sds\" (UniqueName: \"kubernetes.io/projected/36d22a9d-a20f-460d-b030-63792ec5da2c-kube-api-access-v7sds\") pod \"36d22a9d-a20f-460d-b030-63792ec5da2c\" (UID: \"36d22a9d-a20f-460d-b030-63792ec5da2c\") " Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.451080 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36d22a9d-a20f-460d-b030-63792ec5da2c-host" (OuterVolumeSpecName: "host") pod "36d22a9d-a20f-460d-b030-63792ec5da2c" (UID: "36d22a9d-a20f-460d-b030-63792ec5da2c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.451563 4679 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d22a9d-a20f-460d-b030-63792ec5da2c-host\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.473878 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d22a9d-a20f-460d-b030-63792ec5da2c-kube-api-access-v7sds" (OuterVolumeSpecName: "kube-api-access-v7sds") pod "36d22a9d-a20f-460d-b030-63792ec5da2c" (UID: "36d22a9d-a20f-460d-b030-63792ec5da2c"). InnerVolumeSpecName "kube-api-access-v7sds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.552888 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7sds\" (UniqueName: \"kubernetes.io/projected/36d22a9d-a20f-460d-b030-63792ec5da2c-kube-api-access-v7sds\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:37 crc kubenswrapper[4679]: I1202 11:18:37.730422 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f896p"] Dec 02 11:18:37 crc kubenswrapper[4679]: W1202 11:18:37.732935 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ae41fa0_44ae_475f_b9d0_0cfe9363ea0e.slice/crio-2f25302757b834b036f7beeee1f63d871bc9732f3a8d69e82f2ad825265de1ce WatchSource:0}: Error finding container 2f25302757b834b036f7beeee1f63d871bc9732f3a8d69e82f2ad825265de1ce: Status 404 returned error can't find the container with id 2f25302757b834b036f7beeee1f63d871bc9732f3a8d69e82f2ad825265de1ce Dec 02 11:18:38 crc kubenswrapper[4679]: I1202 11:18:38.219893 4679 scope.go:117] "RemoveContainer" containerID="ac5e6552023906620c4859520b8e6f014bd832302d9afadb24e5f14dc3e67370" Dec 02 11:18:38 crc kubenswrapper[4679]: I1202 11:18:38.219944 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/crc-debug-c79gt" Dec 02 11:18:38 crc kubenswrapper[4679]: I1202 11:18:38.221756 4679 generic.go:334] "Generic (PLEG): container finished" podID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerID="65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a" exitCode=0 Dec 02 11:18:38 crc kubenswrapper[4679]: I1202 11:18:38.221854 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerDied","Data":"65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a"} Dec 02 11:18:38 crc kubenswrapper[4679]: I1202 11:18:38.221930 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerStarted","Data":"2f25302757b834b036f7beeee1f63d871bc9732f3a8d69e82f2ad825265de1ce"} Dec 02 11:18:38 crc kubenswrapper[4679]: I1202 11:18:38.928260 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d22a9d-a20f-460d-b030-63792ec5da2c" path="/var/lib/kubelet/pods/36d22a9d-a20f-460d-b030-63792ec5da2c/volumes" Dec 02 11:18:39 crc kubenswrapper[4679]: I1202 11:18:39.232572 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerStarted","Data":"e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690"} Dec 02 11:18:40 crc kubenswrapper[4679]: I1202 11:18:40.243366 4679 generic.go:334] "Generic (PLEG): container finished" podID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerID="e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690" exitCode=0 Dec 02 11:18:40 crc kubenswrapper[4679]: I1202 11:18:40.243436 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerDied","Data":"e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690"} Dec 02 11:18:41 crc kubenswrapper[4679]: I1202 11:18:41.252874 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerStarted","Data":"5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75"} Dec 02 11:18:41 crc kubenswrapper[4679]: I1202 11:18:41.271993 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f896p" podStartSLOduration=2.634282297 podStartE2EDuration="5.271975136s" podCreationTimestamp="2025-12-02 11:18:36 +0000 UTC" firstStartedPulling="2025-12-02 11:18:38.223562039 +0000 UTC m=+3571.553700899" lastFinishedPulling="2025-12-02 11:18:40.861254878 +0000 UTC m=+3574.191393738" observedRunningTime="2025-12-02 11:18:41.268461586 +0000 UTC m=+3574.598600446" watchObservedRunningTime="2025-12-02 11:18:41.271975136 +0000 UTC m=+3574.602113996" Dec 02 11:18:46 crc kubenswrapper[4679]: I1202 11:18:46.931634 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:18:46 crc kubenswrapper[4679]: I1202 11:18:46.931987 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:18:47 crc kubenswrapper[4679]: I1202 11:18:47.228629 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:47 crc kubenswrapper[4679]: I1202 11:18:47.230321 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:47 crc kubenswrapper[4679]: I1202 11:18:47.291710 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:47 crc kubenswrapper[4679]: I1202 11:18:47.363221 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:47 crc kubenswrapper[4679]: I1202 11:18:47.542547 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f896p"] Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.319511 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f896p" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="registry-server" containerID="cri-o://5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75" gracePeriod=2 Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.752491 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.883852 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-catalog-content\") pod \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.884180 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-utilities\") pod \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.884400 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjmwb\" (UniqueName: \"kubernetes.io/projected/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-kube-api-access-vjmwb\") pod \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\" (UID: \"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e\") " Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.886523 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-utilities" (OuterVolumeSpecName: "utilities") pod "5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" (UID: "5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.892175 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-kube-api-access-vjmwb" (OuterVolumeSpecName: "kube-api-access-vjmwb") pod "5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" (UID: "5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e"). InnerVolumeSpecName "kube-api-access-vjmwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.938275 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" (UID: "5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.986809 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.986850 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:49 crc kubenswrapper[4679]: I1202 11:18:49.986863 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjmwb\" (UniqueName: \"kubernetes.io/projected/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e-kube-api-access-vjmwb\") on node \"crc\" DevicePath \"\"" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.331658 4679 generic.go:334] "Generic (PLEG): container finished" podID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerID="5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75" exitCode=0 Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.331709 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerDied","Data":"5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75"} Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.331747 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f896p" event={"ID":"5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e","Type":"ContainerDied","Data":"2f25302757b834b036f7beeee1f63d871bc9732f3a8d69e82f2ad825265de1ce"} Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.331768 4679 scope.go:117] "RemoveContainer" containerID="5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.333120 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f896p" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.365117 4679 scope.go:117] "RemoveContainer" containerID="e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.370361 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f896p"] Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.380502 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f896p"] Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.386635 4679 scope.go:117] "RemoveContainer" containerID="65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.430232 4679 scope.go:117] "RemoveContainer" containerID="5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75" Dec 02 11:18:50 crc kubenswrapper[4679]: E1202 11:18:50.430606 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75\": container with ID starting with 5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75 not found: ID does not exist" containerID="5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.430634 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75"} err="failed to get container status \"5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75\": rpc error: code = NotFound desc = could not find container \"5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75\": container with ID starting with 5035c1b416fbea526d36bb8bdde9115fe5bb4ae0fa65d5f9cd493dbbcc928e75 not found: ID does not exist" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.430654 4679 scope.go:117] "RemoveContainer" containerID="e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690" Dec 02 11:18:50 crc kubenswrapper[4679]: E1202 11:18:50.430847 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690\": container with ID starting with e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690 not found: ID does not exist" containerID="e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.430866 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690"} err="failed to get container status \"e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690\": rpc error: code = NotFound desc = could not find container \"e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690\": container with ID starting with e7ffec39676e479ee6178940026a500073622e5bb3c9e34953a6fe286a5c7690 not found: ID does not exist" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.430878 4679 scope.go:117] "RemoveContainer" containerID="65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a" Dec 02 11:18:50 crc kubenswrapper[4679]: E1202 11:18:50.431075 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a\": container with ID starting with 65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a not found: ID does not exist" containerID="65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.431098 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a"} err="failed to get container status \"65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a\": rpc error: code = NotFound desc = could not find container \"65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a\": container with ID starting with 65879da4a2897e899530c0081cb4938b3f5f2f33b248fc3de6a92d56d93f102a not found: ID does not exist" Dec 02 11:18:50 crc kubenswrapper[4679]: I1202 11:18:50.921087 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" path="/var/lib/kubelet/pods/5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e/volumes" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.079584 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd9984b94-spfgc_028dd5f7-9b68-45af-8515-abef1cc975d0/barbican-api/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.133108 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd9984b94-spfgc_028dd5f7-9b68-45af-8515-abef1cc975d0/barbican-api-log/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.264683 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78df7f5646-r7nmv_3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f/barbican-keystone-listener/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.369364 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78df7f5646-r7nmv_3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f/barbican-keystone-listener-log/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.473524 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9dcfb6847-f2pxh_5b25066f-17f7-42c3-8448-a25ff170388a/barbican-worker/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.529227 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9dcfb6847-f2pxh_5b25066f-17f7-42c3-8448-a25ff170388a/barbican-worker-log/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.598817 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw_dd93c8e7-6ac8-4061-a426-43433fef0376/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.759757 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/ceilometer-central-agent/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.836704 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/ceilometer-notification-agent/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.847765 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/proxy-httpd/0.log" Dec 02 11:18:52 crc kubenswrapper[4679]: I1202 11:18:52.931939 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/sg-core/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.050641 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5f8a8618-590b-498d-ae0f-ab9f9f83b054/cinder-api/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.052228 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5f8a8618-590b-498d-ae0f-ab9f9f83b054/cinder-api-log/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.242413 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_247813fb-2103-4014-b0a3-c5be6310a181/cinder-scheduler/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.279514 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_247813fb-2103-4014-b0a3-c5be6310a181/probe/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.396268 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-thzwv_c6a42827-37a6-46ee-bc6a-925b0434d21a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.495989 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ff56m_9947dc7b-205c-463b-98d9-07759646f56f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.614169 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-b4d9r_c203b3f2-0c27-4421-9912-9ac96318189c/init/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.773238 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-b4d9r_c203b3f2-0c27-4421-9912-9ac96318189c/init/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.809431 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-b4d9r_c203b3f2-0c27-4421-9912-9ac96318189c/dnsmasq-dns/0.log" Dec 02 11:18:53 crc kubenswrapper[4679]: I1202 11:18:53.828100 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2_2119b0d5-f21e-4694-9fc8-7bf3bed24ecc/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.170237 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4386b4f2-7580-4c18-ae6a-078f8de0edc5/glance-httpd/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.195670 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4386b4f2-7580-4c18-ae6a-078f8de0edc5/glance-log/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.358950 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_daa65ea1-139c-49c6-9f88-36efbb79d52c/glance-httpd/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.414194 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_daa65ea1-139c-49c6-9f88-36efbb79d52c/glance-log/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.477543 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6fd99d7d5b-sb96h_4dd62da9-d481-4fc3-bbe6-108a54d98154/horizon/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.679460 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm_ca2fc1cb-e547-47c7-b27a-7de8429d2032/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.872839 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6fd99d7d5b-sb96h_4dd62da9-d481-4fc3-bbe6-108a54d98154/horizon-log/0.log" Dec 02 11:18:54 crc kubenswrapper[4679]: I1202 11:18:54.913891 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nr9dr_645181f0-3043-4bad-908e-8d57baa1d9b9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:55 crc kubenswrapper[4679]: I1202 11:18:55.126157 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29411221-9cggb_dea51cbc-b2de-4d00-a13e-fb2e1c66b605/keystone-cron/0.log" Dec 02 11:18:55 crc kubenswrapper[4679]: I1202 11:18:55.184913 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6558c68964-dhdww_6691f3fc-9b58-4cc5-af4e-b4aaf10036e9/keystone-api/0.log" Dec 02 11:18:55 crc kubenswrapper[4679]: I1202 11:18:55.310042 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_84bc340a-3e0b-4ac5-94a7-07156dbcbbf0/kube-state-metrics/0.log" Dec 02 11:18:55 crc kubenswrapper[4679]: I1202 11:18:55.434011 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lhf76_1202831e-ebbe-4507-b89a-b3e9de9d64e6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:55 crc kubenswrapper[4679]: I1202 11:18:55.836249 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6dbdc9c48f-dtqv6_eecb1bea-c78b-4888-b9a8-e868ab82a9db/neutron-httpd/0.log" Dec 02 11:18:55 crc kubenswrapper[4679]: I1202 11:18:55.871290 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6dbdc9c48f-dtqv6_eecb1bea-c78b-4888-b9a8-e868ab82a9db/neutron-api/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.063678 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r_7764127f-88d3-4c9e-9eb2-9a6c37585ca3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.468935 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_759b2c49-d8b4-4f12-b60d-3ad7765f8aa3/nova-cell0-conductor-conductor/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.509558 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_81f22355-2ad7-4fd6-bc88-779cd704b038/nova-api-log/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.677557 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_81f22355-2ad7-4fd6-bc88-779cd704b038/nova-api-api/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.741776 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_95edebcc-4231-4f68-be9d-7e04a0d15e16/nova-cell1-conductor-conductor/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.826323 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_30a95365-228b-4db8-bcae-4957f4a4ee87/nova-cell1-novncproxy-novncproxy/0.log" Dec 02 11:18:56 crc kubenswrapper[4679]: I1202 11:18:56.987639 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-nvlmk_eae6d050-4b37-4c45-a0bd-4b930852eecf/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:57 crc kubenswrapper[4679]: I1202 11:18:57.108734 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8d049f08-8029-4ef7-9383-34925ba3d48b/nova-metadata-log/0.log" Dec 02 11:18:57 crc kubenswrapper[4679]: I1202 11:18:57.409688 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d731d782-b47c-4e03-8191-47ae5d52a555/nova-scheduler-scheduler/0.log" Dec 02 11:18:57 crc kubenswrapper[4679]: I1202 11:18:57.413348 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_30ab65a3-79d4-4eb7-a6b3-c759422ddffc/mysql-bootstrap/0.log" Dec 02 11:18:57 crc kubenswrapper[4679]: I1202 11:18:57.586128 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_30ab65a3-79d4-4eb7-a6b3-c759422ddffc/mysql-bootstrap/0.log" Dec 02 11:18:57 crc kubenswrapper[4679]: I1202 11:18:57.631956 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_30ab65a3-79d4-4eb7-a6b3-c759422ddffc/galera/0.log" Dec 02 11:18:57 crc kubenswrapper[4679]: I1202 11:18:57.957577 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e1f514bf-70bf-484c-b485-7e3d46a5c139/mysql-bootstrap/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.064256 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e1f514bf-70bf-484c-b485-7e3d46a5c139/mysql-bootstrap/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.101840 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e1f514bf-70bf-484c-b485-7e3d46a5c139/galera/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.230749 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8d049f08-8029-4ef7-9383-34925ba3d48b/nova-metadata-metadata/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.231165 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c0d970b7-84af-4667-8b4e-6b2d4b800923/openstackclient/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.405900 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kg24f_142f095e-2fa0-478d-a611-50aad287d2c6/openstack-network-exporter/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.455183 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovsdb-server-init/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.691524 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovsdb-server-init/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.734559 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovs-vswitchd/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.763849 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovsdb-server/0.log" Dec 02 11:18:58 crc kubenswrapper[4679]: I1202 11:18:58.930368 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p5dsw_9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb/ovn-controller/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.060878 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gzdgr_7e6fa9dc-0546-45fd-847c-29266ce64dfa/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.198645 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ae2e8d7b-ca00-404b-a11b-4ea010116548/ovn-northd/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.209751 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ae2e8d7b-ca00-404b-a11b-4ea010116548/openstack-network-exporter/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.349961 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_71d9df94-9e39-409b-aa9e-e13a8f8f9ea2/openstack-network-exporter/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.408671 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_71d9df94-9e39-409b-aa9e-e13a8f8f9ea2/ovsdbserver-nb/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.544173 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_285ae6d4-044a-4720-a495-8ca9eca431f3/openstack-network-exporter/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.634735 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_285ae6d4-044a-4720-a495-8ca9eca431f3/ovsdbserver-sb/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.804561 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-947c4b5c6-7fq4t_19f15bf3-d397-4937-bcd3-c61999ce8dae/placement-api/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.854846 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-947c4b5c6-7fq4t_19f15bf3-d397-4937-bcd3-c61999ce8dae/placement-log/0.log" Dec 02 11:18:59 crc kubenswrapper[4679]: I1202 11:18:59.916856 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17e32beb-4297-4540-a48b-b1d3d4a32ee2/setup-container/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.087169 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17e32beb-4297-4540-a48b-b1d3d4a32ee2/setup-container/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.178889 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17e32beb-4297-4540-a48b-b1d3d4a32ee2/rabbitmq/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.214061 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_561abb24-2326-4992-b863-6065fbdcb959/setup-container/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.340831 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_561abb24-2326-4992-b863-6065fbdcb959/setup-container/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.391932 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_561abb24-2326-4992-b863-6065fbdcb959/rabbitmq/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.459993 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj_0430fbc5-3d73-4885-8670-61dd256714be/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.602500 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-bzv4b_e5d79c8f-5612-4597-ba1a-51108bf84927/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.715467 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6_52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.803390 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-7dlsf_2fd59ebd-72da-4ad7-93b5-7f51f3a3304c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:19:00 crc kubenswrapper[4679]: I1202 11:19:00.966524 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-jtmsz_8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7/ssh-known-hosts-edpm-deployment/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.178376 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8d8658b7-4bkfp_8d434783-1fe8-4ba4-8195-282fa213066d/proxy-server/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.232364 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8d8658b7-4bkfp_8d434783-1fe8-4ba4-8195-282fa213066d/proxy-httpd/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.376793 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-2wk8m_714d01a2-8f79-41d8-a16a-47c35a69fe8b/swift-ring-rebalance/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.647454 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-reaper/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.694622 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-auditor/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.713281 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-replicator/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.830464 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-server/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.874862 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-auditor/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.911084 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-server/0.log" Dec 02 11:19:01 crc kubenswrapper[4679]: I1202 11:19:01.989970 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-replicator/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.077604 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-updater/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.114435 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-expirer/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.166537 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-auditor/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.265838 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-replicator/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.271676 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-server/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.342804 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-updater/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.422902 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/rsync/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.505849 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/swift-recon-cron/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.618666 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6_d79ba6bd-2e41-485d-a11f-5b0bdb5500c8/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.727872 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_564a5130-acc1-40e0-814b-2e9bfd2d967d/tempest-tests-tempest-tests-runner/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.833261 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d17cc270-8f47-4ea2-ba89-6bb3011ed620/test-operator-logs-container/0.log" Dec 02 11:19:02 crc kubenswrapper[4679]: I1202 11:19:02.952004 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5nrms_1420f674-925e-449b-8e18-0e97225f3958/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.169820 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tbvh8"] Dec 02 11:19:05 crc kubenswrapper[4679]: E1202 11:19:05.170510 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="extract-content" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.170522 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="extract-content" Dec 02 11:19:05 crc kubenswrapper[4679]: E1202 11:19:05.170545 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="registry-server" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.170551 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="registry-server" Dec 02 11:19:05 crc kubenswrapper[4679]: E1202 11:19:05.170587 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="extract-utilities" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.170595 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="extract-utilities" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.170776 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae41fa0-44ae-475f-b9d0-0cfe9363ea0e" containerName="registry-server" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.172100 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.185420 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tbvh8"] Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.257962 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfkwq\" (UniqueName: \"kubernetes.io/projected/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-kube-api-access-bfkwq\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.258278 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-utilities\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.258459 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-catalog-content\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.361477 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfkwq\" (UniqueName: \"kubernetes.io/projected/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-kube-api-access-bfkwq\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.361584 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-utilities\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.361646 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-catalog-content\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.362956 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-utilities\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.364984 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-catalog-content\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.381999 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfkwq\" (UniqueName: \"kubernetes.io/projected/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-kube-api-access-bfkwq\") pod \"community-operators-tbvh8\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:05 crc kubenswrapper[4679]: I1202 11:19:05.497133 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:06 crc kubenswrapper[4679]: I1202 11:19:06.293862 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tbvh8"] Dec 02 11:19:06 crc kubenswrapper[4679]: I1202 11:19:06.489294 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerStarted","Data":"5d93fc3f691edc32c5b68e9f0aa552ba6fe190efb8a90849ca7104d5a3cb7770"} Dec 02 11:19:07 crc kubenswrapper[4679]: I1202 11:19:07.499840 4679 generic.go:334] "Generic (PLEG): container finished" podID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerID="b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878" exitCode=0 Dec 02 11:19:07 crc kubenswrapper[4679]: I1202 11:19:07.499913 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerDied","Data":"b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878"} Dec 02 11:19:08 crc kubenswrapper[4679]: I1202 11:19:08.510031 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerStarted","Data":"c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927"} Dec 02 11:19:09 crc kubenswrapper[4679]: I1202 11:19:09.521389 4679 generic.go:334] "Generic (PLEG): container finished" podID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerID="c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927" exitCode=0 Dec 02 11:19:09 crc kubenswrapper[4679]: I1202 11:19:09.521460 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerDied","Data":"c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927"} Dec 02 11:19:10 crc kubenswrapper[4679]: I1202 11:19:10.586602 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerStarted","Data":"c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357"} Dec 02 11:19:10 crc kubenswrapper[4679]: I1202 11:19:10.626379 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tbvh8" podStartSLOduration=3.065314076 podStartE2EDuration="5.626360071s" podCreationTimestamp="2025-12-02 11:19:05 +0000 UTC" firstStartedPulling="2025-12-02 11:19:07.503656216 +0000 UTC m=+3600.833795076" lastFinishedPulling="2025-12-02 11:19:10.064702211 +0000 UTC m=+3603.394841071" observedRunningTime="2025-12-02 11:19:10.616687596 +0000 UTC m=+3603.946826456" watchObservedRunningTime="2025-12-02 11:19:10.626360071 +0000 UTC m=+3603.956498931" Dec 02 11:19:13 crc kubenswrapper[4679]: I1202 11:19:13.736702 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_afdf0aae-7bec-4c47-820f-5a970c264069/memcached/0.log" Dec 02 11:19:15 crc kubenswrapper[4679]: I1202 11:19:15.497889 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:15 crc kubenswrapper[4679]: I1202 11:19:15.498164 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:15 crc kubenswrapper[4679]: I1202 11:19:15.548089 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:15 crc kubenswrapper[4679]: I1202 11:19:15.690885 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:15 crc kubenswrapper[4679]: I1202 11:19:15.782337 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tbvh8"] Dec 02 11:19:16 crc kubenswrapper[4679]: I1202 11:19:16.931556 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:19:16 crc kubenswrapper[4679]: I1202 11:19:16.931629 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:19:17 crc kubenswrapper[4679]: I1202 11:19:17.658222 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tbvh8" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="registry-server" containerID="cri-o://c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357" gracePeriod=2 Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.092907 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.229378 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-utilities\") pod \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.229448 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfkwq\" (UniqueName: \"kubernetes.io/projected/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-kube-api-access-bfkwq\") pod \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.229560 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-catalog-content\") pod \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\" (UID: \"e7aaf1d0-13eb-4399-8d48-827b815b8d6a\") " Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.230560 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-utilities" (OuterVolumeSpecName: "utilities") pod "e7aaf1d0-13eb-4399-8d48-827b815b8d6a" (UID: "e7aaf1d0-13eb-4399-8d48-827b815b8d6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.232154 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.235219 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-kube-api-access-bfkwq" (OuterVolumeSpecName: "kube-api-access-bfkwq") pod "e7aaf1d0-13eb-4399-8d48-827b815b8d6a" (UID: "e7aaf1d0-13eb-4399-8d48-827b815b8d6a"). InnerVolumeSpecName "kube-api-access-bfkwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.282936 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7aaf1d0-13eb-4399-8d48-827b815b8d6a" (UID: "e7aaf1d0-13eb-4399-8d48-827b815b8d6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.334296 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfkwq\" (UniqueName: \"kubernetes.io/projected/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-kube-api-access-bfkwq\") on node \"crc\" DevicePath \"\"" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.334357 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7aaf1d0-13eb-4399-8d48-827b815b8d6a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.671535 4679 generic.go:334] "Generic (PLEG): container finished" podID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerID="c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357" exitCode=0 Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.671626 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerDied","Data":"c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357"} Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.671912 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tbvh8" event={"ID":"e7aaf1d0-13eb-4399-8d48-827b815b8d6a","Type":"ContainerDied","Data":"5d93fc3f691edc32c5b68e9f0aa552ba6fe190efb8a90849ca7104d5a3cb7770"} Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.671941 4679 scope.go:117] "RemoveContainer" containerID="c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.671649 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tbvh8" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.703865 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tbvh8"] Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.710537 4679 scope.go:117] "RemoveContainer" containerID="c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.712911 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tbvh8"] Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.735399 4679 scope.go:117] "RemoveContainer" containerID="b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.778050 4679 scope.go:117] "RemoveContainer" containerID="c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357" Dec 02 11:19:18 crc kubenswrapper[4679]: E1202 11:19:18.778609 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357\": container with ID starting with c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357 not found: ID does not exist" containerID="c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.778651 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357"} err="failed to get container status \"c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357\": rpc error: code = NotFound desc = could not find container \"c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357\": container with ID starting with c414bcda3d32b6eb452d99f8ff1d9025e33250b3ea11d97e81ea9b5e3b18c357 not found: ID does not exist" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.778677 4679 scope.go:117] "RemoveContainer" containerID="c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927" Dec 02 11:19:18 crc kubenswrapper[4679]: E1202 11:19:18.779232 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927\": container with ID starting with c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927 not found: ID does not exist" containerID="c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.779256 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927"} err="failed to get container status \"c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927\": rpc error: code = NotFound desc = could not find container \"c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927\": container with ID starting with c71f86cdfec140d3babff500112aae4706d93c4a9cffe06d2ae63ccaee66b927 not found: ID does not exist" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.779269 4679 scope.go:117] "RemoveContainer" containerID="b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878" Dec 02 11:19:18 crc kubenswrapper[4679]: E1202 11:19:18.779612 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878\": container with ID starting with b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878 not found: ID does not exist" containerID="b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.779636 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878"} err="failed to get container status \"b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878\": rpc error: code = NotFound desc = could not find container \"b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878\": container with ID starting with b752138d030f84caef5ce830b13d9ad4e851b5432fcf36ef91866f143b71a878 not found: ID does not exist" Dec 02 11:19:18 crc kubenswrapper[4679]: I1202 11:19:18.922879 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" path="/var/lib/kubelet/pods/e7aaf1d0-13eb-4399-8d48-827b815b8d6a/volumes" Dec 02 11:19:28 crc kubenswrapper[4679]: I1202 11:19:28.860588 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jmj2d_48f2b621-5d9b-4213-8681-8107d1fc04a4/manager/0.log" Dec 02 11:19:28 crc kubenswrapper[4679]: I1202 11:19:28.920927 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jmj2d_48f2b621-5d9b-4213-8681-8107d1fc04a4/kube-rbac-proxy/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.049262 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-snss6_7938c8f2-53f6-4143-997d-a44ea54c9fc2/kube-rbac-proxy/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.114585 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-snss6_7938c8f2-53f6-4143-997d-a44ea54c9fc2/manager/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.222076 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/util/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.383754 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/util/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.406145 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/pull/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.406720 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/pull/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.653403 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/util/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.654334 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/extract/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.657980 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/pull/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.842444 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-59k6f_85ed4b5c-eda1-481a-84ae-b97d01a9b9eb/kube-rbac-proxy/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.846168 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-59k6f_85ed4b5c-eda1-481a-84ae-b97d01a9b9eb/manager/0.log" Dec 02 11:19:29 crc kubenswrapper[4679]: I1202 11:19:29.939005 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-442vr_61bf694c-dbb6-442c-aaa0-3c13ff7efc97/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.075379 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jxm5w_f9cf4107-9e23-4246-a8b7-b3bbda0cc461/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.107388 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-442vr_61bf694c-dbb6-442c-aaa0-3c13ff7efc97/manager/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.128482 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jxm5w_f9cf4107-9e23-4246-a8b7-b3bbda0cc461/manager/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.250272 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bnz2w_80f86afb-0d79-4d1c-a0fd-b1d573d3c471/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.296780 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bnz2w_80f86afb-0d79-4d1c-a0fd-b1d573d3c471/manager/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.439182 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-ghgsb_a0088242-8397-41e7-96b8-d971f307eefc/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.498941 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-5zzmd_5ee03695-d4bb-4b77-84d1-84190ae307d6/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.671484 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-5zzmd_5ee03695-d4bb-4b77-84d1-84190ae307d6/manager/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.680022 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-ghgsb_a0088242-8397-41e7-96b8-d971f307eefc/manager/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.726608 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-sxgbw_ebbb7257-823e-49a1-af4c-703b979de700/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.882519 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-tp7zw_1cf6ea83-714b-4a9d-8a5e-708293a3e9e6/kube-rbac-proxy/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.898763 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-tp7zw_1cf6ea83-714b-4a9d-8a5e-708293a3e9e6/manager/0.log" Dec 02 11:19:30 crc kubenswrapper[4679]: I1202 11:19:30.987648 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-sxgbw_ebbb7257-823e-49a1-af4c-703b979de700/manager/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.095219 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-w74h9_e334fbef-5aab-4cd6-83a1-084160e32d90/kube-rbac-proxy/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.165813 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-w74h9_e334fbef-5aab-4cd6-83a1-084160e32d90/manager/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.287493 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-48bsh_e23de616-947b-4ff1-b05b-522a1c39cc7b/kube-rbac-proxy/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.322064 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-48bsh_e23de616-947b-4ff1-b05b-522a1c39cc7b/manager/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.387603 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-628rp_9dc9ab59-e368-452c-9ce2-e4d028fd35a1/kube-rbac-proxy/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.545578 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-628rp_9dc9ab59-e368-452c-9ce2-e4d028fd35a1/manager/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.609762 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-nm552_1b3b0e6e-eb2f-40fc-a77f-c8889816dee3/manager/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.617720 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-nm552_1b3b0e6e-eb2f-40fc-a77f-c8889816dee3/kube-rbac-proxy/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.768128 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh_b255bbc9-0452-4445-9470-8bd7ad85257a/kube-rbac-proxy/0.log" Dec 02 11:19:31 crc kubenswrapper[4679]: I1202 11:19:31.796236 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh_b255bbc9-0452-4445-9470-8bd7ad85257a/manager/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.210778 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-655b495c79-ktwld_7b731918-9630-4abe-a4ce-ce8a65167295/operator/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.236544 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wh4rf_7fa3fa68-e237-45d4-b1b4-7422e84f46ef/registry-server/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.405433 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xfz4c_3fa279c2-e6ef-4131-ab48-28306a24000e/kube-rbac-proxy/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.505958 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xfz4c_3fa279c2-e6ef-4131-ab48-28306a24000e/manager/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.637899 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9flx2_c26eed77-aa87-4b6b-9f81-aefe3e5ed786/kube-rbac-proxy/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.683026 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9flx2_c26eed77-aa87-4b6b-9f81-aefe3e5ed786/manager/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.751995 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lxjdn_0a119e38-647a-42ab-823f-e5ecea4afbf0/operator/0.log" Dec 02 11:19:32 crc kubenswrapper[4679]: I1202 11:19:32.899693 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nw7fm_a877b082-6f7d-4f85-b290-356fe7c8bb5f/kube-rbac-proxy/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.013046 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nw7fm_a877b082-6f7d-4f85-b290-356fe7c8bb5f/manager/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.095620 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7bc5fbbb46-rqv9x_6edbf2ab-201d-4a7f-902a-7b74566b5a36/manager/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.097188 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-ps9nb_921e63bc-e66b-48d5-a22c-010c65a0e1b0/kube-rbac-proxy/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.230808 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-ps9nb_921e63bc-e66b-48d5-a22c-010c65a0e1b0/manager/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.248834 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-b74hj_6ac36ac3-f2ba-40be-9544-9897a27ddaf4/kube-rbac-proxy/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.266460 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-b74hj_6ac36ac3-f2ba-40be-9544-9897a27ddaf4/manager/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.426684 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7kl62_f48fce46-9ae8-4154-b67b-3edd56267209/kube-rbac-proxy/0.log" Dec 02 11:19:33 crc kubenswrapper[4679]: I1202 11:19:33.428690 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7kl62_f48fce46-9ae8-4154-b67b-3edd56267209/manager/0.log" Dec 02 11:19:46 crc kubenswrapper[4679]: I1202 11:19:46.931575 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:19:46 crc kubenswrapper[4679]: I1202 11:19:46.932049 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:19:46 crc kubenswrapper[4679]: I1202 11:19:46.932082 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 11:19:46 crc kubenswrapper[4679]: I1202 11:19:46.932755 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68ca6b38ae7ce6a11d6cd6e7bfb27a17db2519f3b4ecdb2f0b95937e41fad9d0"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 11:19:46 crc kubenswrapper[4679]: I1202 11:19:46.932804 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://68ca6b38ae7ce6a11d6cd6e7bfb27a17db2519f3b4ecdb2f0b95937e41fad9d0" gracePeriod=600 Dec 02 11:19:47 crc kubenswrapper[4679]: I1202 11:19:47.991648 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="68ca6b38ae7ce6a11d6cd6e7bfb27a17db2519f3b4ecdb2f0b95937e41fad9d0" exitCode=0 Dec 02 11:19:47 crc kubenswrapper[4679]: I1202 11:19:47.992297 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"68ca6b38ae7ce6a11d6cd6e7bfb27a17db2519f3b4ecdb2f0b95937e41fad9d0"} Dec 02 11:19:47 crc kubenswrapper[4679]: I1202 11:19:47.992363 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8"} Dec 02 11:19:47 crc kubenswrapper[4679]: I1202 11:19:47.992384 4679 scope.go:117] "RemoveContainer" containerID="2672aad4ec41d0a5ceb96e73a7ed2687859e83cb36a74cc395193d363c6bd62a" Dec 02 11:19:51 crc kubenswrapper[4679]: I1202 11:19:51.078618 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ffc27_6e31c209-d885-4235-87a7-972b045a6c72/control-plane-machine-set-operator/0.log" Dec 02 11:19:51 crc kubenswrapper[4679]: I1202 11:19:51.285495 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z95rb_e222153a-8cab-4ef0-8227-13be178f3f82/machine-api-operator/0.log" Dec 02 11:19:51 crc kubenswrapper[4679]: I1202 11:19:51.290041 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z95rb_e222153a-8cab-4ef0-8227-13be178f3f82/kube-rbac-proxy/0.log" Dec 02 11:20:03 crc kubenswrapper[4679]: I1202 11:20:03.083036 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-rnwm9_1da80a36-7d00-4055-ba3e-1e84a58e2748/cert-manager-controller/0.log" Dec 02 11:20:03 crc kubenswrapper[4679]: I1202 11:20:03.213458 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dhkj7_227ec57c-bcbe-4725-8726-3411f96b09ac/cert-manager-cainjector/0.log" Dec 02 11:20:03 crc kubenswrapper[4679]: I1202 11:20:03.261963 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-lhvkg_f91be00f-75e8-4a8b-aacc-d881b6af57a8/cert-manager-webhook/0.log" Dec 02 11:20:15 crc kubenswrapper[4679]: I1202 11:20:15.238097 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-r2xnw_932b0be3-5659-4b5b-a667-cd03f56c8f98/nmstate-console-plugin/0.log" Dec 02 11:20:15 crc kubenswrapper[4679]: I1202 11:20:15.436797 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-t6qm2_49ed79ab-5582-4f7e-bfd9-18263fa9ccc2/nmstate-handler/0.log" Dec 02 11:20:15 crc kubenswrapper[4679]: I1202 11:20:15.491716 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9w2kz_1daefa02-7428-44b6-8da5-f1ecf0d5c12e/nmstate-metrics/0.log" Dec 02 11:20:15 crc kubenswrapper[4679]: I1202 11:20:15.506325 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9w2kz_1daefa02-7428-44b6-8da5-f1ecf0d5c12e/kube-rbac-proxy/0.log" Dec 02 11:20:15 crc kubenswrapper[4679]: I1202 11:20:15.688088 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-hnqnh_0cdf81d6-53ba-4fae-95ff-6d0d28b2384e/nmstate-operator/0.log" Dec 02 11:20:15 crc kubenswrapper[4679]: I1202 11:20:15.736555 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-fmbsj_338504ed-81c0-4c88-8d99-b1ffc9cd7bcb/nmstate-webhook/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.234423 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-ws8bf_d2302ab5-3f22-4f2e-804f-a9e1109446a6/kube-rbac-proxy/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.323540 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-ws8bf_d2302ab5-3f22-4f2e-804f-a9e1109446a6/controller/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.461893 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.743361 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.759378 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.776583 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.785391 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.979807 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.984070 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.984363 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:20:30 crc kubenswrapper[4679]: I1202 11:20:30.988986 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.175539 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.175553 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.181117 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/controller/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.196891 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.355470 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/frr-metrics/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.369504 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/kube-rbac-proxy/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.394589 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/kube-rbac-proxy-frr/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.640095 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/reloader/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.642374 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-zcqd5_280f2b1e-4e5a-481d-ad2f-4536d832697a/frr-k8s-webhook-server/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.849631 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8555c5bdfd-j4hqj_06025ad8-fd55-4fe1-95bd-08a0c83168e0/manager/0.log" Dec 02 11:20:31 crc kubenswrapper[4679]: I1202 11:20:31.988633 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-898774cd7-nvg9q_4086bfe3-e82b-46c4-87f6-837a42348472/webhook-server/0.log" Dec 02 11:20:32 crc kubenswrapper[4679]: I1202 11:20:32.119886 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qgdwn_c42448eb-c465-4172-9830-37000ff7e881/kube-rbac-proxy/0.log" Dec 02 11:20:32 crc kubenswrapper[4679]: I1202 11:20:32.704659 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qgdwn_c42448eb-c465-4172-9830-37000ff7e881/speaker/0.log" Dec 02 11:20:32 crc kubenswrapper[4679]: I1202 11:20:32.758635 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/frr/0.log" Dec 02 11:20:44 crc kubenswrapper[4679]: I1202 11:20:44.977661 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/util/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.149088 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/util/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.183279 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/pull/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.183406 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/pull/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.380118 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/pull/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.381848 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/util/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.404843 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/extract/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.525006 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/util/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.702728 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/util/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.748199 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/pull/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.755154 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/pull/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.912971 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/util/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.932101 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/pull/0.log" Dec 02 11:20:45 crc kubenswrapper[4679]: I1202 11:20:45.956178 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/extract/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.119691 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-utilities/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.299042 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-content/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.333868 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-utilities/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.343489 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-content/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.483117 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-utilities/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.508670 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-content/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.756816 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-utilities/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.917530 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-content/0.log" Dec 02 11:20:46 crc kubenswrapper[4679]: I1202 11:20:46.970479 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/registry-server/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.002021 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-utilities/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.022911 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-content/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.163538 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-content/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.165082 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-utilities/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.423072 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-72flf_de76d395-f3a6-474d-8ab7-f31c46710a0d/marketplace-operator/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.581757 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/registry-server/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.590396 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-utilities/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.750449 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-content/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.753838 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-utilities/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.814809 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-content/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.971484 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-utilities/0.log" Dec 02 11:20:47 crc kubenswrapper[4679]: I1202 11:20:47.978211 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-content/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.125829 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/registry-server/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.185793 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-utilities/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.378708 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-utilities/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.404556 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-content/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.417872 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-content/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.626256 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-content/0.log" Dec 02 11:20:48 crc kubenswrapper[4679]: I1202 11:20:48.638718 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-utilities/0.log" Dec 02 11:20:49 crc kubenswrapper[4679]: I1202 11:20:49.031033 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/registry-server/0.log" Dec 02 11:22:16 crc kubenswrapper[4679]: I1202 11:22:16.931508 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:22:16 crc kubenswrapper[4679]: I1202 11:22:16.932034 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:22:27 crc kubenswrapper[4679]: I1202 11:22:27.552512 4679 generic.go:334] "Generic (PLEG): container finished" podID="2d48b322-087a-4173-8110-262905e0d358" containerID="5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0" exitCode=0 Dec 02 11:22:27 crc kubenswrapper[4679]: I1202 11:22:27.552607 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-529wn/must-gather-x86c5" event={"ID":"2d48b322-087a-4173-8110-262905e0d358","Type":"ContainerDied","Data":"5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0"} Dec 02 11:22:27 crc kubenswrapper[4679]: I1202 11:22:27.553692 4679 scope.go:117] "RemoveContainer" containerID="5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0" Dec 02 11:22:27 crc kubenswrapper[4679]: I1202 11:22:27.652005 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-529wn_must-gather-x86c5_2d48b322-087a-4173-8110-262905e0d358/gather/0.log" Dec 02 11:22:34 crc kubenswrapper[4679]: I1202 11:22:34.689778 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-529wn/must-gather-x86c5"] Dec 02 11:22:34 crc kubenswrapper[4679]: I1202 11:22:34.692854 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-529wn/must-gather-x86c5" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="copy" containerID="cri-o://796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f" gracePeriod=2 Dec 02 11:22:34 crc kubenswrapper[4679]: I1202 11:22:34.701528 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-529wn/must-gather-x86c5"] Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.151784 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-529wn_must-gather-x86c5_2d48b322-087a-4173-8110-262905e0d358/copy/0.log" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.152754 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.270875 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxglw\" (UniqueName: \"kubernetes.io/projected/2d48b322-087a-4173-8110-262905e0d358-kube-api-access-lxglw\") pod \"2d48b322-087a-4173-8110-262905e0d358\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.271180 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d48b322-087a-4173-8110-262905e0d358-must-gather-output\") pod \"2d48b322-087a-4173-8110-262905e0d358\" (UID: \"2d48b322-087a-4173-8110-262905e0d358\") " Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.278881 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d48b322-087a-4173-8110-262905e0d358-kube-api-access-lxglw" (OuterVolumeSpecName: "kube-api-access-lxglw") pod "2d48b322-087a-4173-8110-262905e0d358" (UID: "2d48b322-087a-4173-8110-262905e0d358"). InnerVolumeSpecName "kube-api-access-lxglw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.380127 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxglw\" (UniqueName: \"kubernetes.io/projected/2d48b322-087a-4173-8110-262905e0d358-kube-api-access-lxglw\") on node \"crc\" DevicePath \"\"" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.455866 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d48b322-087a-4173-8110-262905e0d358-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2d48b322-087a-4173-8110-262905e0d358" (UID: "2d48b322-087a-4173-8110-262905e0d358"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.481410 4679 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d48b322-087a-4173-8110-262905e0d358-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.645057 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-529wn_must-gather-x86c5_2d48b322-087a-4173-8110-262905e0d358/copy/0.log" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.645782 4679 generic.go:334] "Generic (PLEG): container finished" podID="2d48b322-087a-4173-8110-262905e0d358" containerID="796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f" exitCode=143 Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.645826 4679 scope.go:117] "RemoveContainer" containerID="796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.645919 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-529wn/must-gather-x86c5" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.692969 4679 scope.go:117] "RemoveContainer" containerID="5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.749208 4679 scope.go:117] "RemoveContainer" containerID="796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f" Dec 02 11:22:35 crc kubenswrapper[4679]: E1202 11:22:35.749676 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f\": container with ID starting with 796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f not found: ID does not exist" containerID="796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.749727 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f"} err="failed to get container status \"796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f\": rpc error: code = NotFound desc = could not find container \"796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f\": container with ID starting with 796cd3ad5ce8a13a3da11e85fb80b7f3e42b4137ffa634526f93c34857da398f not found: ID does not exist" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.749761 4679 scope.go:117] "RemoveContainer" containerID="5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0" Dec 02 11:22:35 crc kubenswrapper[4679]: E1202 11:22:35.750205 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0\": container with ID starting with 5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0 not found: ID does not exist" containerID="5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0" Dec 02 11:22:35 crc kubenswrapper[4679]: I1202 11:22:35.750231 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0"} err="failed to get container status \"5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0\": rpc error: code = NotFound desc = could not find container \"5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0\": container with ID starting with 5a42fee4367ede39e90c20924db03916cf8f1e51bb66d9f63df15417d8ea4ec0 not found: ID does not exist" Dec 02 11:22:36 crc kubenswrapper[4679]: I1202 11:22:36.931160 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d48b322-087a-4173-8110-262905e0d358" path="/var/lib/kubelet/pods/2d48b322-087a-4173-8110-262905e0d358/volumes" Dec 02 11:22:46 crc kubenswrapper[4679]: I1202 11:22:46.931553 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:22:46 crc kubenswrapper[4679]: I1202 11:22:46.932428 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.820297 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p9kzs"] Dec 02 11:23:10 crc kubenswrapper[4679]: E1202 11:23:10.824464 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="registry-server" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.824683 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="registry-server" Dec 02 11:23:10 crc kubenswrapper[4679]: E1202 11:23:10.824782 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="copy" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.824859 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="copy" Dec 02 11:23:10 crc kubenswrapper[4679]: E1202 11:23:10.824950 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="gather" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.825026 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="gather" Dec 02 11:23:10 crc kubenswrapper[4679]: E1202 11:23:10.825118 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="extract-content" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.825207 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="extract-content" Dec 02 11:23:10 crc kubenswrapper[4679]: E1202 11:23:10.825298 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="extract-utilities" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.825431 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="extract-utilities" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.825839 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7aaf1d0-13eb-4399-8d48-827b815b8d6a" containerName="registry-server" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.825981 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="copy" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.826108 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d48b322-087a-4173-8110-262905e0d358" containerName="gather" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.828708 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.834579 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9kzs"] Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.901177 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8csd8\" (UniqueName: \"kubernetes.io/projected/2db13405-601e-4e6e-8202-2c4093be09e5-kube-api-access-8csd8\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.901299 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-catalog-content\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:10 crc kubenswrapper[4679]: I1202 11:23:10.901394 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-utilities\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.002995 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8csd8\" (UniqueName: \"kubernetes.io/projected/2db13405-601e-4e6e-8202-2c4093be09e5-kube-api-access-8csd8\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.003373 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-catalog-content\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.003517 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-utilities\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.003922 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-catalog-content\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.003979 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-utilities\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.027394 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8csd8\" (UniqueName: \"kubernetes.io/projected/2db13405-601e-4e6e-8202-2c4093be09e5-kube-api-access-8csd8\") pod \"redhat-marketplace-p9kzs\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.147077 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:11 crc kubenswrapper[4679]: I1202 11:23:11.603880 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9kzs"] Dec 02 11:23:12 crc kubenswrapper[4679]: I1202 11:23:12.035605 4679 generic.go:334] "Generic (PLEG): container finished" podID="2db13405-601e-4e6e-8202-2c4093be09e5" containerID="58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc" exitCode=0 Dec 02 11:23:12 crc kubenswrapper[4679]: I1202 11:23:12.035686 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerDied","Data":"58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc"} Dec 02 11:23:12 crc kubenswrapper[4679]: I1202 11:23:12.035734 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerStarted","Data":"dbb6b7966c9a378c55179ceb8ad783107a96fa980e67bf99c1f8e663cfe9706e"} Dec 02 11:23:12 crc kubenswrapper[4679]: I1202 11:23:12.038220 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 11:23:13 crc kubenswrapper[4679]: I1202 11:23:13.045853 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerStarted","Data":"7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c"} Dec 02 11:23:14 crc kubenswrapper[4679]: I1202 11:23:14.059916 4679 generic.go:334] "Generic (PLEG): container finished" podID="2db13405-601e-4e6e-8202-2c4093be09e5" containerID="7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c" exitCode=0 Dec 02 11:23:14 crc kubenswrapper[4679]: I1202 11:23:14.059968 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerDied","Data":"7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c"} Dec 02 11:23:15 crc kubenswrapper[4679]: I1202 11:23:15.096783 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerStarted","Data":"b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428"} Dec 02 11:23:15 crc kubenswrapper[4679]: I1202 11:23:15.119065 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p9kzs" podStartSLOduration=2.516167473 podStartE2EDuration="5.119019763s" podCreationTimestamp="2025-12-02 11:23:10 +0000 UTC" firstStartedPulling="2025-12-02 11:23:12.037706419 +0000 UTC m=+3845.367845319" lastFinishedPulling="2025-12-02 11:23:14.640558739 +0000 UTC m=+3847.970697609" observedRunningTime="2025-12-02 11:23:15.113893867 +0000 UTC m=+3848.444032737" watchObservedRunningTime="2025-12-02 11:23:15.119019763 +0000 UTC m=+3848.449158623" Dec 02 11:23:16 crc kubenswrapper[4679]: I1202 11:23:16.932205 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:23:16 crc kubenswrapper[4679]: I1202 11:23:16.932658 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:23:16 crc kubenswrapper[4679]: I1202 11:23:16.932714 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 11:23:16 crc kubenswrapper[4679]: I1202 11:23:16.933572 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 11:23:16 crc kubenswrapper[4679]: I1202 11:23:16.933680 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" gracePeriod=600 Dec 02 11:23:17 crc kubenswrapper[4679]: E1202 11:23:17.064701 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:23:17 crc kubenswrapper[4679]: I1202 11:23:17.119201 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" exitCode=0 Dec 02 11:23:17 crc kubenswrapper[4679]: I1202 11:23:17.119247 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8"} Dec 02 11:23:17 crc kubenswrapper[4679]: I1202 11:23:17.119287 4679 scope.go:117] "RemoveContainer" containerID="68ca6b38ae7ce6a11d6cd6e7bfb27a17db2519f3b4ecdb2f0b95937e41fad9d0" Dec 02 11:23:17 crc kubenswrapper[4679]: I1202 11:23:17.120085 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:23:17 crc kubenswrapper[4679]: E1202 11:23:17.120441 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:23:21 crc kubenswrapper[4679]: I1202 11:23:21.147419 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:21 crc kubenswrapper[4679]: I1202 11:23:21.147952 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:21 crc kubenswrapper[4679]: I1202 11:23:21.240221 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:21 crc kubenswrapper[4679]: I1202 11:23:21.321502 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:21 crc kubenswrapper[4679]: I1202 11:23:21.482315 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9kzs"] Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.181166 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p9kzs" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="registry-server" containerID="cri-o://b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428" gracePeriod=2 Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.660839 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.769268 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-utilities\") pod \"2db13405-601e-4e6e-8202-2c4093be09e5\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.769412 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8csd8\" (UniqueName: \"kubernetes.io/projected/2db13405-601e-4e6e-8202-2c4093be09e5-kube-api-access-8csd8\") pod \"2db13405-601e-4e6e-8202-2c4093be09e5\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.769483 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-catalog-content\") pod \"2db13405-601e-4e6e-8202-2c4093be09e5\" (UID: \"2db13405-601e-4e6e-8202-2c4093be09e5\") " Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.770248 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-utilities" (OuterVolumeSpecName: "utilities") pod "2db13405-601e-4e6e-8202-2c4093be09e5" (UID: "2db13405-601e-4e6e-8202-2c4093be09e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.776769 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db13405-601e-4e6e-8202-2c4093be09e5-kube-api-access-8csd8" (OuterVolumeSpecName: "kube-api-access-8csd8") pod "2db13405-601e-4e6e-8202-2c4093be09e5" (UID: "2db13405-601e-4e6e-8202-2c4093be09e5"). InnerVolumeSpecName "kube-api-access-8csd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.795788 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2db13405-601e-4e6e-8202-2c4093be09e5" (UID: "2db13405-601e-4e6e-8202-2c4093be09e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.871045 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.871080 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2db13405-601e-4e6e-8202-2c4093be09e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:23:23 crc kubenswrapper[4679]: I1202 11:23:23.871091 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8csd8\" (UniqueName: \"kubernetes.io/projected/2db13405-601e-4e6e-8202-2c4093be09e5-kube-api-access-8csd8\") on node \"crc\" DevicePath \"\"" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.192706 4679 generic.go:334] "Generic (PLEG): container finished" podID="2db13405-601e-4e6e-8202-2c4093be09e5" containerID="b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428" exitCode=0 Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.192948 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerDied","Data":"b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428"} Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.194110 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9kzs" event={"ID":"2db13405-601e-4e6e-8202-2c4093be09e5","Type":"ContainerDied","Data":"dbb6b7966c9a378c55179ceb8ad783107a96fa980e67bf99c1f8e663cfe9706e"} Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.194295 4679 scope.go:117] "RemoveContainer" containerID="b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.193068 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9kzs" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.229200 4679 scope.go:117] "RemoveContainer" containerID="7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.253612 4679 scope.go:117] "RemoveContainer" containerID="58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.255491 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9kzs"] Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.290937 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9kzs"] Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.310570 4679 scope.go:117] "RemoveContainer" containerID="b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428" Dec 02 11:23:24 crc kubenswrapper[4679]: E1202 11:23:24.314528 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428\": container with ID starting with b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428 not found: ID does not exist" containerID="b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.314588 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428"} err="failed to get container status \"b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428\": rpc error: code = NotFound desc = could not find container \"b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428\": container with ID starting with b4fb07aefd7da96542f0bc747bf133ea21cce1dc009c10b92222de973ab4e428 not found: ID does not exist" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.314619 4679 scope.go:117] "RemoveContainer" containerID="7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c" Dec 02 11:23:24 crc kubenswrapper[4679]: E1202 11:23:24.319562 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c\": container with ID starting with 7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c not found: ID does not exist" containerID="7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.319621 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c"} err="failed to get container status \"7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c\": rpc error: code = NotFound desc = could not find container \"7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c\": container with ID starting with 7cca894e3e0c24929d34d31c6fb9609e3a6437b419644180a5123e35bc10227c not found: ID does not exist" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.319652 4679 scope.go:117] "RemoveContainer" containerID="58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc" Dec 02 11:23:24 crc kubenswrapper[4679]: E1202 11:23:24.323422 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc\": container with ID starting with 58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc not found: ID does not exist" containerID="58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.323471 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc"} err="failed to get container status \"58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc\": rpc error: code = NotFound desc = could not find container \"58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc\": container with ID starting with 58c272984b3958fc7eed8c4dc0948d0aafe2a294a274d7f6f1a9cf11581fc3cc not found: ID does not exist" Dec 02 11:23:24 crc kubenswrapper[4679]: I1202 11:23:24.920669 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" path="/var/lib/kubelet/pods/2db13405-601e-4e6e-8202-2c4093be09e5/volumes" Dec 02 11:23:30 crc kubenswrapper[4679]: I1202 11:23:30.909595 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:23:30 crc kubenswrapper[4679]: E1202 11:23:30.910509 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:23:45 crc kubenswrapper[4679]: I1202 11:23:45.909780 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:23:45 crc kubenswrapper[4679]: E1202 11:23:45.910441 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:24:00 crc kubenswrapper[4679]: I1202 11:24:00.908892 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:24:00 crc kubenswrapper[4679]: E1202 11:24:00.909937 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:24:15 crc kubenswrapper[4679]: I1202 11:24:15.909003 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:24:15 crc kubenswrapper[4679]: E1202 11:24:15.910010 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:24:23 crc kubenswrapper[4679]: I1202 11:24:23.742212 4679 scope.go:117] "RemoveContainer" containerID="47958b45aabc226c8d4431a181963d1eda9445c5a7d92444d455858369c86c35" Dec 02 11:24:26 crc kubenswrapper[4679]: I1202 11:24:26.914473 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:24:26 crc kubenswrapper[4679]: E1202 11:24:26.915320 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:24:38 crc kubenswrapper[4679]: I1202 11:24:38.909229 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:24:38 crc kubenswrapper[4679]: E1202 11:24:38.912126 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:24:50 crc kubenswrapper[4679]: I1202 11:24:50.908842 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:24:50 crc kubenswrapper[4679]: E1202 11:24:50.909528 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:25:04 crc kubenswrapper[4679]: I1202 11:25:04.908938 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:25:04 crc kubenswrapper[4679]: E1202 11:25:04.910765 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.048178 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kpfxs"] Dec 02 11:25:11 crc kubenswrapper[4679]: E1202 11:25:11.049095 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="extract-utilities" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.049110 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="extract-utilities" Dec 02 11:25:11 crc kubenswrapper[4679]: E1202 11:25:11.049127 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="extract-content" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.049135 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="extract-content" Dec 02 11:25:11 crc kubenswrapper[4679]: E1202 11:25:11.049176 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="registry-server" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.049185 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="registry-server" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.049437 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db13405-601e-4e6e-8202-2c4093be09e5" containerName="registry-server" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.051260 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.062616 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kpfxs"] Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.172675 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-utilities\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.173067 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s7jn\" (UniqueName: \"kubernetes.io/projected/22251d6a-3596-4842-8809-bb48a0c14d54-kube-api-access-5s7jn\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.173195 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-catalog-content\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.274856 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-utilities\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.274960 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s7jn\" (UniqueName: \"kubernetes.io/projected/22251d6a-3596-4842-8809-bb48a0c14d54-kube-api-access-5s7jn\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.274994 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-catalog-content\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.275821 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-utilities\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.275914 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-catalog-content\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.308995 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s7jn\" (UniqueName: \"kubernetes.io/projected/22251d6a-3596-4842-8809-bb48a0c14d54-kube-api-access-5s7jn\") pod \"redhat-operators-kpfxs\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.385620 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:11 crc kubenswrapper[4679]: I1202 11:25:11.869587 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kpfxs"] Dec 02 11:25:12 crc kubenswrapper[4679]: I1202 11:25:12.316317 4679 generic.go:334] "Generic (PLEG): container finished" podID="22251d6a-3596-4842-8809-bb48a0c14d54" containerID="8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201" exitCode=0 Dec 02 11:25:12 crc kubenswrapper[4679]: I1202 11:25:12.316355 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpfxs" event={"ID":"22251d6a-3596-4842-8809-bb48a0c14d54","Type":"ContainerDied","Data":"8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201"} Dec 02 11:25:12 crc kubenswrapper[4679]: I1202 11:25:12.316380 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpfxs" event={"ID":"22251d6a-3596-4842-8809-bb48a0c14d54","Type":"ContainerStarted","Data":"4c162d0d45f7e373ecd6112aa4297a724c0b2e49bdf7c2813a436f0dcda6d615"} Dec 02 11:25:14 crc kubenswrapper[4679]: I1202 11:25:14.337802 4679 generic.go:334] "Generic (PLEG): container finished" podID="22251d6a-3596-4842-8809-bb48a0c14d54" containerID="e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f" exitCode=0 Dec 02 11:25:14 crc kubenswrapper[4679]: I1202 11:25:14.337896 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpfxs" event={"ID":"22251d6a-3596-4842-8809-bb48a0c14d54","Type":"ContainerDied","Data":"e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f"} Dec 02 11:25:16 crc kubenswrapper[4679]: I1202 11:25:16.367519 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpfxs" event={"ID":"22251d6a-3596-4842-8809-bb48a0c14d54","Type":"ContainerStarted","Data":"793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2"} Dec 02 11:25:16 crc kubenswrapper[4679]: I1202 11:25:16.413598 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kpfxs" podStartSLOduration=2.546157488 podStartE2EDuration="5.413566846s" podCreationTimestamp="2025-12-02 11:25:11 +0000 UTC" firstStartedPulling="2025-12-02 11:25:12.318651021 +0000 UTC m=+3965.648789881" lastFinishedPulling="2025-12-02 11:25:15.186060379 +0000 UTC m=+3968.516199239" observedRunningTime="2025-12-02 11:25:16.40284419 +0000 UTC m=+3969.732983120" watchObservedRunningTime="2025-12-02 11:25:16.413566846 +0000 UTC m=+3969.743705746" Dec 02 11:25:18 crc kubenswrapper[4679]: I1202 11:25:18.909449 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:25:18 crc kubenswrapper[4679]: E1202 11:25:18.910033 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:25:21 crc kubenswrapper[4679]: I1202 11:25:21.387657 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:21 crc kubenswrapper[4679]: I1202 11:25:21.388088 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:21 crc kubenswrapper[4679]: I1202 11:25:21.452098 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:21 crc kubenswrapper[4679]: I1202 11:25:21.500717 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:21 crc kubenswrapper[4679]: I1202 11:25:21.687780 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kpfxs"] Dec 02 11:25:22 crc kubenswrapper[4679]: I1202 11:25:22.954261 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s9jkv/must-gather-hbkdr"] Dec 02 11:25:22 crc kubenswrapper[4679]: I1202 11:25:22.956107 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:22 crc kubenswrapper[4679]: I1202 11:25:22.958619 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-s9jkv"/"default-dockercfg-l46zq" Dec 02 11:25:22 crc kubenswrapper[4679]: I1202 11:25:22.962195 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s9jkv"/"kube-root-ca.crt" Dec 02 11:25:22 crc kubenswrapper[4679]: I1202 11:25:22.962225 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s9jkv"/"openshift-service-ca.crt" Dec 02 11:25:22 crc kubenswrapper[4679]: I1202 11:25:22.972326 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s9jkv/must-gather-hbkdr"] Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.132249 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1d5ac99-bc19-482e-b627-b5072881dfad-must-gather-output\") pod \"must-gather-hbkdr\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.132344 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6vvg\" (UniqueName: \"kubernetes.io/projected/e1d5ac99-bc19-482e-b627-b5072881dfad-kube-api-access-p6vvg\") pod \"must-gather-hbkdr\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.235114 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1d5ac99-bc19-482e-b627-b5072881dfad-must-gather-output\") pod \"must-gather-hbkdr\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.235184 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6vvg\" (UniqueName: \"kubernetes.io/projected/e1d5ac99-bc19-482e-b627-b5072881dfad-kube-api-access-p6vvg\") pod \"must-gather-hbkdr\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.235564 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1d5ac99-bc19-482e-b627-b5072881dfad-must-gather-output\") pod \"must-gather-hbkdr\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.254463 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6vvg\" (UniqueName: \"kubernetes.io/projected/e1d5ac99-bc19-482e-b627-b5072881dfad-kube-api-access-p6vvg\") pod \"must-gather-hbkdr\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.280476 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.435800 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kpfxs" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="registry-server" containerID="cri-o://793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2" gracePeriod=2 Dec 02 11:25:23 crc kubenswrapper[4679]: I1202 11:25:23.803277 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s9jkv/must-gather-hbkdr"] Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.025323 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.157138 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-catalog-content\") pod \"22251d6a-3596-4842-8809-bb48a0c14d54\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.157701 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-utilities\") pod \"22251d6a-3596-4842-8809-bb48a0c14d54\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.157727 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s7jn\" (UniqueName: \"kubernetes.io/projected/22251d6a-3596-4842-8809-bb48a0c14d54-kube-api-access-5s7jn\") pod \"22251d6a-3596-4842-8809-bb48a0c14d54\" (UID: \"22251d6a-3596-4842-8809-bb48a0c14d54\") " Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.159124 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-utilities" (OuterVolumeSpecName: "utilities") pod "22251d6a-3596-4842-8809-bb48a0c14d54" (UID: "22251d6a-3596-4842-8809-bb48a0c14d54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.165442 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22251d6a-3596-4842-8809-bb48a0c14d54-kube-api-access-5s7jn" (OuterVolumeSpecName: "kube-api-access-5s7jn") pod "22251d6a-3596-4842-8809-bb48a0c14d54" (UID: "22251d6a-3596-4842-8809-bb48a0c14d54"). InnerVolumeSpecName "kube-api-access-5s7jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.260530 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.260566 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s7jn\" (UniqueName: \"kubernetes.io/projected/22251d6a-3596-4842-8809-bb48a0c14d54-kube-api-access-5s7jn\") on node \"crc\" DevicePath \"\"" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.446556 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" event={"ID":"e1d5ac99-bc19-482e-b627-b5072881dfad","Type":"ContainerStarted","Data":"721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340"} Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.446610 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" event={"ID":"e1d5ac99-bc19-482e-b627-b5072881dfad","Type":"ContainerStarted","Data":"4435b1c1a53b375d36ffff08cf7eaf9896440e020605c28d227c6c90787bd30a"} Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.448756 4679 generic.go:334] "Generic (PLEG): container finished" podID="22251d6a-3596-4842-8809-bb48a0c14d54" containerID="793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2" exitCode=0 Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.448792 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpfxs" event={"ID":"22251d6a-3596-4842-8809-bb48a0c14d54","Type":"ContainerDied","Data":"793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2"} Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.448814 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kpfxs" event={"ID":"22251d6a-3596-4842-8809-bb48a0c14d54","Type":"ContainerDied","Data":"4c162d0d45f7e373ecd6112aa4297a724c0b2e49bdf7c2813a436f0dcda6d615"} Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.448834 4679 scope.go:117] "RemoveContainer" containerID="793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.449001 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kpfxs" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.474931 4679 scope.go:117] "RemoveContainer" containerID="e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.511690 4679 scope.go:117] "RemoveContainer" containerID="8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.549528 4679 scope.go:117] "RemoveContainer" containerID="793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2" Dec 02 11:25:24 crc kubenswrapper[4679]: E1202 11:25:24.550142 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2\": container with ID starting with 793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2 not found: ID does not exist" containerID="793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.550183 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2"} err="failed to get container status \"793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2\": rpc error: code = NotFound desc = could not find container \"793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2\": container with ID starting with 793d88c1e3384e45edda7582491fc81b52704bb74b147b8031bc6290f456cef2 not found: ID does not exist" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.550208 4679 scope.go:117] "RemoveContainer" containerID="e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f" Dec 02 11:25:24 crc kubenswrapper[4679]: E1202 11:25:24.550719 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f\": container with ID starting with e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f not found: ID does not exist" containerID="e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.550773 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f"} err="failed to get container status \"e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f\": rpc error: code = NotFound desc = could not find container \"e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f\": container with ID starting with e5a18ae187867aa16b336ab8095891f2d3566ff0a88a7d49824139ca48366b4f not found: ID does not exist" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.550806 4679 scope.go:117] "RemoveContainer" containerID="8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201" Dec 02 11:25:24 crc kubenswrapper[4679]: E1202 11:25:24.551219 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201\": container with ID starting with 8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201 not found: ID does not exist" containerID="8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201" Dec 02 11:25:24 crc kubenswrapper[4679]: I1202 11:25:24.551264 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201"} err="failed to get container status \"8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201\": rpc error: code = NotFound desc = could not find container \"8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201\": container with ID starting with 8f2a2457dca5f33259794bd065c29d1312be5e6c7efb77e63bd7c6be4370f201 not found: ID does not exist" Dec 02 11:25:25 crc kubenswrapper[4679]: I1202 11:25:25.403199 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22251d6a-3596-4842-8809-bb48a0c14d54" (UID: "22251d6a-3596-4842-8809-bb48a0c14d54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:25:25 crc kubenswrapper[4679]: I1202 11:25:25.457888 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" event={"ID":"e1d5ac99-bc19-482e-b627-b5072881dfad","Type":"ContainerStarted","Data":"3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea"} Dec 02 11:25:25 crc kubenswrapper[4679]: I1202 11:25:25.479914 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" podStartSLOduration=3.479893802 podStartE2EDuration="3.479893802s" podCreationTimestamp="2025-12-02 11:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 11:25:25.472809361 +0000 UTC m=+3978.802948231" watchObservedRunningTime="2025-12-02 11:25:25.479893802 +0000 UTC m=+3978.810032662" Dec 02 11:25:25 crc kubenswrapper[4679]: I1202 11:25:25.485555 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22251d6a-3596-4842-8809-bb48a0c14d54-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:25:25 crc kubenswrapper[4679]: I1202 11:25:25.700131 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kpfxs"] Dec 02 11:25:25 crc kubenswrapper[4679]: I1202 11:25:25.711938 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kpfxs"] Dec 02 11:25:26 crc kubenswrapper[4679]: I1202 11:25:26.921277 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" path="/var/lib/kubelet/pods/22251d6a-3596-4842-8809-bb48a0c14d54/volumes" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.018487 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-4glgp"] Dec 02 11:25:28 crc kubenswrapper[4679]: E1202 11:25:28.019212 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="extract-utilities" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.019228 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="extract-utilities" Dec 02 11:25:28 crc kubenswrapper[4679]: E1202 11:25:28.019252 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="extract-content" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.019261 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="extract-content" Dec 02 11:25:28 crc kubenswrapper[4679]: E1202 11:25:28.019320 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="registry-server" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.019330 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="registry-server" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.019568 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="22251d6a-3596-4842-8809-bb48a0c14d54" containerName="registry-server" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.020406 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.132992 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0bbf1981-7c85-46af-bd62-7220c2ecd98b-host\") pod \"crc-debug-4glgp\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.133336 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvl9p\" (UniqueName: \"kubernetes.io/projected/0bbf1981-7c85-46af-bd62-7220c2ecd98b-kube-api-access-cvl9p\") pod \"crc-debug-4glgp\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.235918 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvl9p\" (UniqueName: \"kubernetes.io/projected/0bbf1981-7c85-46af-bd62-7220c2ecd98b-kube-api-access-cvl9p\") pod \"crc-debug-4glgp\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.236120 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0bbf1981-7c85-46af-bd62-7220c2ecd98b-host\") pod \"crc-debug-4glgp\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.236270 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0bbf1981-7c85-46af-bd62-7220c2ecd98b-host\") pod \"crc-debug-4glgp\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.256086 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvl9p\" (UniqueName: \"kubernetes.io/projected/0bbf1981-7c85-46af-bd62-7220c2ecd98b-kube-api-access-cvl9p\") pod \"crc-debug-4glgp\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.342903 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:25:28 crc kubenswrapper[4679]: I1202 11:25:28.503507 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" event={"ID":"0bbf1981-7c85-46af-bd62-7220c2ecd98b","Type":"ContainerStarted","Data":"2ab52662f1a881b3f73c60be845fe07f3175918fc7620e7b3cc2fbe8f03d8200"} Dec 02 11:25:29 crc kubenswrapper[4679]: I1202 11:25:29.512872 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" event={"ID":"0bbf1981-7c85-46af-bd62-7220c2ecd98b","Type":"ContainerStarted","Data":"d92fefe3bb1f71c6d247e276764700af6f076811179648e6f61ece326f58cc0c"} Dec 02 11:25:29 crc kubenswrapper[4679]: I1202 11:25:29.530226 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" podStartSLOduration=1.530210246 podStartE2EDuration="1.530210246s" podCreationTimestamp="2025-12-02 11:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 11:25:29.525424259 +0000 UTC m=+3982.855563119" watchObservedRunningTime="2025-12-02 11:25:29.530210246 +0000 UTC m=+3982.860349106" Dec 02 11:25:30 crc kubenswrapper[4679]: I1202 11:25:30.908810 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:25:30 crc kubenswrapper[4679]: E1202 11:25:30.909413 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:25:44 crc kubenswrapper[4679]: I1202 11:25:44.909558 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:25:44 crc kubenswrapper[4679]: E1202 11:25:44.910208 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:25:56 crc kubenswrapper[4679]: I1202 11:25:56.916584 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:25:56 crc kubenswrapper[4679]: E1202 11:25:56.918370 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:26:03 crc kubenswrapper[4679]: I1202 11:26:03.814459 4679 generic.go:334] "Generic (PLEG): container finished" podID="0bbf1981-7c85-46af-bd62-7220c2ecd98b" containerID="d92fefe3bb1f71c6d247e276764700af6f076811179648e6f61ece326f58cc0c" exitCode=0 Dec 02 11:26:03 crc kubenswrapper[4679]: I1202 11:26:03.815411 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" event={"ID":"0bbf1981-7c85-46af-bd62-7220c2ecd98b","Type":"ContainerDied","Data":"d92fefe3bb1f71c6d247e276764700af6f076811179648e6f61ece326f58cc0c"} Dec 02 11:26:04 crc kubenswrapper[4679]: I1202 11:26:04.939696 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:26:04 crc kubenswrapper[4679]: I1202 11:26:04.974469 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-4glgp"] Dec 02 11:26:04 crc kubenswrapper[4679]: I1202 11:26:04.983262 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-4glgp"] Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.078646 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvl9p\" (UniqueName: \"kubernetes.io/projected/0bbf1981-7c85-46af-bd62-7220c2ecd98b-kube-api-access-cvl9p\") pod \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.079429 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0bbf1981-7c85-46af-bd62-7220c2ecd98b-host\") pod \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\" (UID: \"0bbf1981-7c85-46af-bd62-7220c2ecd98b\") " Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.079503 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bbf1981-7c85-46af-bd62-7220c2ecd98b-host" (OuterVolumeSpecName: "host") pod "0bbf1981-7c85-46af-bd62-7220c2ecd98b" (UID: "0bbf1981-7c85-46af-bd62-7220c2ecd98b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.080393 4679 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0bbf1981-7c85-46af-bd62-7220c2ecd98b-host\") on node \"crc\" DevicePath \"\"" Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.088552 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bbf1981-7c85-46af-bd62-7220c2ecd98b-kube-api-access-cvl9p" (OuterVolumeSpecName: "kube-api-access-cvl9p") pod "0bbf1981-7c85-46af-bd62-7220c2ecd98b" (UID: "0bbf1981-7c85-46af-bd62-7220c2ecd98b"). InnerVolumeSpecName "kube-api-access-cvl9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.183202 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvl9p\" (UniqueName: \"kubernetes.io/projected/0bbf1981-7c85-46af-bd62-7220c2ecd98b-kube-api-access-cvl9p\") on node \"crc\" DevicePath \"\"" Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.832130 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ab52662f1a881b3f73c60be845fe07f3175918fc7620e7b3cc2fbe8f03d8200" Dec 02 11:26:05 crc kubenswrapper[4679]: I1202 11:26:05.832224 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-4glgp" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.227206 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-286r7"] Dec 02 11:26:06 crc kubenswrapper[4679]: E1202 11:26:06.228495 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bbf1981-7c85-46af-bd62-7220c2ecd98b" containerName="container-00" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.228608 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bbf1981-7c85-46af-bd62-7220c2ecd98b" containerName="container-00" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.228942 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bbf1981-7c85-46af-bd62-7220c2ecd98b" containerName="container-00" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.230058 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.403816 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65891f44-3766-41c8-9485-d9d6ee9d045a-host\") pod \"crc-debug-286r7\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.404214 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg75p\" (UniqueName: \"kubernetes.io/projected/65891f44-3766-41c8-9485-d9d6ee9d045a-kube-api-access-jg75p\") pod \"crc-debug-286r7\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.505740 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg75p\" (UniqueName: \"kubernetes.io/projected/65891f44-3766-41c8-9485-d9d6ee9d045a-kube-api-access-jg75p\") pod \"crc-debug-286r7\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.505925 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65891f44-3766-41c8-9485-d9d6ee9d045a-host\") pod \"crc-debug-286r7\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.506020 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65891f44-3766-41c8-9485-d9d6ee9d045a-host\") pod \"crc-debug-286r7\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.527495 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg75p\" (UniqueName: \"kubernetes.io/projected/65891f44-3766-41c8-9485-d9d6ee9d045a-kube-api-access-jg75p\") pod \"crc-debug-286r7\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.557135 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.842373 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-286r7" event={"ID":"65891f44-3766-41c8-9485-d9d6ee9d045a","Type":"ContainerStarted","Data":"48626f710891507314d5fa36e7f5096df92661247b8ab235adf117c6f678eec8"} Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.842417 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-286r7" event={"ID":"65891f44-3766-41c8-9485-d9d6ee9d045a","Type":"ContainerStarted","Data":"15052bf353653e4b7322439b0cb469ec1e9d4c85e85b84658d46eaaab2141894"} Dec 02 11:26:06 crc kubenswrapper[4679]: I1202 11:26:06.920888 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bbf1981-7c85-46af-bd62-7220c2ecd98b" path="/var/lib/kubelet/pods/0bbf1981-7c85-46af-bd62-7220c2ecd98b/volumes" Dec 02 11:26:07 crc kubenswrapper[4679]: I1202 11:26:07.342210 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-286r7"] Dec 02 11:26:07 crc kubenswrapper[4679]: I1202 11:26:07.344925 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-286r7"] Dec 02 11:26:07 crc kubenswrapper[4679]: I1202 11:26:07.853614 4679 generic.go:334] "Generic (PLEG): container finished" podID="65891f44-3766-41c8-9485-d9d6ee9d045a" containerID="48626f710891507314d5fa36e7f5096df92661247b8ab235adf117c6f678eec8" exitCode=0 Dec 02 11:26:07 crc kubenswrapper[4679]: I1202 11:26:07.909431 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:26:07 crc kubenswrapper[4679]: E1202 11:26:07.909786 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:26:07 crc kubenswrapper[4679]: I1202 11:26:07.964353 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.057141 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65891f44-3766-41c8-9485-d9d6ee9d045a-host\") pod \"65891f44-3766-41c8-9485-d9d6ee9d045a\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.057191 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg75p\" (UniqueName: \"kubernetes.io/projected/65891f44-3766-41c8-9485-d9d6ee9d045a-kube-api-access-jg75p\") pod \"65891f44-3766-41c8-9485-d9d6ee9d045a\" (UID: \"65891f44-3766-41c8-9485-d9d6ee9d045a\") " Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.057299 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65891f44-3766-41c8-9485-d9d6ee9d045a-host" (OuterVolumeSpecName: "host") pod "65891f44-3766-41c8-9485-d9d6ee9d045a" (UID: "65891f44-3766-41c8-9485-d9d6ee9d045a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.057655 4679 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/65891f44-3766-41c8-9485-d9d6ee9d045a-host\") on node \"crc\" DevicePath \"\"" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.063128 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65891f44-3766-41c8-9485-d9d6ee9d045a-kube-api-access-jg75p" (OuterVolumeSpecName: "kube-api-access-jg75p") pod "65891f44-3766-41c8-9485-d9d6ee9d045a" (UID: "65891f44-3766-41c8-9485-d9d6ee9d045a"). InnerVolumeSpecName "kube-api-access-jg75p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.159367 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg75p\" (UniqueName: \"kubernetes.io/projected/65891f44-3766-41c8-9485-d9d6ee9d045a-kube-api-access-jg75p\") on node \"crc\" DevicePath \"\"" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.513078 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-w4sgw"] Dec 02 11:26:08 crc kubenswrapper[4679]: E1202 11:26:08.513761 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65891f44-3766-41c8-9485-d9d6ee9d045a" containerName="container-00" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.513775 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="65891f44-3766-41c8-9485-d9d6ee9d045a" containerName="container-00" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.513936 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="65891f44-3766-41c8-9485-d9d6ee9d045a" containerName="container-00" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.514488 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.567557 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6009500f-55a2-4026-8347-9e77410c710c-host\") pod \"crc-debug-w4sgw\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.567694 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2shq\" (UniqueName: \"kubernetes.io/projected/6009500f-55a2-4026-8347-9e77410c710c-kube-api-access-t2shq\") pod \"crc-debug-w4sgw\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.668675 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6009500f-55a2-4026-8347-9e77410c710c-host\") pod \"crc-debug-w4sgw\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.668743 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6009500f-55a2-4026-8347-9e77410c710c-host\") pod \"crc-debug-w4sgw\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.668855 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2shq\" (UniqueName: \"kubernetes.io/projected/6009500f-55a2-4026-8347-9e77410c710c-kube-api-access-t2shq\") pod \"crc-debug-w4sgw\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.684655 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2shq\" (UniqueName: \"kubernetes.io/projected/6009500f-55a2-4026-8347-9e77410c710c-kube-api-access-t2shq\") pod \"crc-debug-w4sgw\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.846300 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.864800 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15052bf353653e4b7322439b0cb469ec1e9d4c85e85b84658d46eaaab2141894" Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.864906 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-286r7" Dec 02 11:26:08 crc kubenswrapper[4679]: W1202 11:26:08.907211 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6009500f_55a2_4026_8347_9e77410c710c.slice/crio-5cb9da9e5dfc01eed5c2f4a7f19c46c74f54b412e2e037ad25547ecf52c5f748 WatchSource:0}: Error finding container 5cb9da9e5dfc01eed5c2f4a7f19c46c74f54b412e2e037ad25547ecf52c5f748: Status 404 returned error can't find the container with id 5cb9da9e5dfc01eed5c2f4a7f19c46c74f54b412e2e037ad25547ecf52c5f748 Dec 02 11:26:08 crc kubenswrapper[4679]: I1202 11:26:08.922400 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65891f44-3766-41c8-9485-d9d6ee9d045a" path="/var/lib/kubelet/pods/65891f44-3766-41c8-9485-d9d6ee9d045a/volumes" Dec 02 11:26:09 crc kubenswrapper[4679]: I1202 11:26:09.874293 4679 generic.go:334] "Generic (PLEG): container finished" podID="6009500f-55a2-4026-8347-9e77410c710c" containerID="c0d757e8b6347d21c7c96c07bbaf790cb263fda4be1cd464293556e0d43ae1b6" exitCode=0 Dec 02 11:26:09 crc kubenswrapper[4679]: I1202 11:26:09.874766 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" event={"ID":"6009500f-55a2-4026-8347-9e77410c710c","Type":"ContainerDied","Data":"c0d757e8b6347d21c7c96c07bbaf790cb263fda4be1cd464293556e0d43ae1b6"} Dec 02 11:26:09 crc kubenswrapper[4679]: I1202 11:26:09.874818 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" event={"ID":"6009500f-55a2-4026-8347-9e77410c710c","Type":"ContainerStarted","Data":"5cb9da9e5dfc01eed5c2f4a7f19c46c74f54b412e2e037ad25547ecf52c5f748"} Dec 02 11:26:09 crc kubenswrapper[4679]: I1202 11:26:09.918406 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-w4sgw"] Dec 02 11:26:09 crc kubenswrapper[4679]: I1202 11:26:09.930263 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s9jkv/crc-debug-w4sgw"] Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.001627 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.112115 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6009500f-55a2-4026-8347-9e77410c710c-host\") pod \"6009500f-55a2-4026-8347-9e77410c710c\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.112380 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6009500f-55a2-4026-8347-9e77410c710c-host" (OuterVolumeSpecName: "host") pod "6009500f-55a2-4026-8347-9e77410c710c" (UID: "6009500f-55a2-4026-8347-9e77410c710c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.112873 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2shq\" (UniqueName: \"kubernetes.io/projected/6009500f-55a2-4026-8347-9e77410c710c-kube-api-access-t2shq\") pod \"6009500f-55a2-4026-8347-9e77410c710c\" (UID: \"6009500f-55a2-4026-8347-9e77410c710c\") " Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.113480 4679 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6009500f-55a2-4026-8347-9e77410c710c-host\") on node \"crc\" DevicePath \"\"" Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.123404 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6009500f-55a2-4026-8347-9e77410c710c-kube-api-access-t2shq" (OuterVolumeSpecName: "kube-api-access-t2shq") pod "6009500f-55a2-4026-8347-9e77410c710c" (UID: "6009500f-55a2-4026-8347-9e77410c710c"). InnerVolumeSpecName "kube-api-access-t2shq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.215737 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2shq\" (UniqueName: \"kubernetes.io/projected/6009500f-55a2-4026-8347-9e77410c710c-kube-api-access-t2shq\") on node \"crc\" DevicePath \"\"" Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.898838 4679 scope.go:117] "RemoveContainer" containerID="c0d757e8b6347d21c7c96c07bbaf790cb263fda4be1cd464293556e0d43ae1b6" Dec 02 11:26:11 crc kubenswrapper[4679]: I1202 11:26:11.899478 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/crc-debug-w4sgw" Dec 02 11:26:12 crc kubenswrapper[4679]: I1202 11:26:12.920285 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6009500f-55a2-4026-8347-9e77410c710c" path="/var/lib/kubelet/pods/6009500f-55a2-4026-8347-9e77410c710c/volumes" Dec 02 11:26:22 crc kubenswrapper[4679]: I1202 11:26:22.909329 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:26:22 crc kubenswrapper[4679]: E1202 11:26:22.910220 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:26:33 crc kubenswrapper[4679]: I1202 11:26:33.909517 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:26:33 crc kubenswrapper[4679]: E1202 11:26:33.910512 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:26:36 crc kubenswrapper[4679]: I1202 11:26:36.926619 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd9984b94-spfgc_028dd5f7-9b68-45af-8515-abef1cc975d0/barbican-api/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.042252 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd9984b94-spfgc_028dd5f7-9b68-45af-8515-abef1cc975d0/barbican-api-log/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.110943 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78df7f5646-r7nmv_3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f/barbican-keystone-listener/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.123521 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78df7f5646-r7nmv_3d67bb3d-3e1d-44d5-9f4d-33a2d39eac9f/barbican-keystone-listener-log/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.626457 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9dcfb6847-f2pxh_5b25066f-17f7-42c3-8448-a25ff170388a/barbican-worker-log/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.687419 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9dcfb6847-f2pxh_5b25066f-17f7-42c3-8448-a25ff170388a/barbican-worker/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.881782 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-v4zzw_dd93c8e7-6ac8-4061-a426-43433fef0376/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.991538 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/ceilometer-notification-agent/0.log" Dec 02 11:26:37 crc kubenswrapper[4679]: I1202 11:26:37.995769 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/ceilometer-central-agent/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.080431 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/proxy-httpd/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.135633 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_17b54231-dc1b-4ccd-8679-b9a697615dbb/sg-core/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.243852 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5f8a8618-590b-498d-ae0f-ab9f9f83b054/cinder-api/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.284766 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_5f8a8618-590b-498d-ae0f-ab9f9f83b054/cinder-api-log/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.460989 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_247813fb-2103-4014-b0a3-c5be6310a181/cinder-scheduler/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.502370 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_247813fb-2103-4014-b0a3-c5be6310a181/probe/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.558080 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-thzwv_c6a42827-37a6-46ee-bc6a-925b0434d21a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.789089 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ff56m_9947dc7b-205c-463b-98d9-07759646f56f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.806856 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-b4d9r_c203b3f2-0c27-4421-9912-9ac96318189c/init/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.918882 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-b4d9r_c203b3f2-0c27-4421-9912-9ac96318189c/init/0.log" Dec 02 11:26:38 crc kubenswrapper[4679]: I1202 11:26:38.968524 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-b4d9r_c203b3f2-0c27-4421-9912-9ac96318189c/dnsmasq-dns/0.log" Dec 02 11:26:39 crc kubenswrapper[4679]: I1202 11:26:39.004831 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-qfvx2_2119b0d5-f21e-4694-9fc8-7bf3bed24ecc/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:39 crc kubenswrapper[4679]: I1202 11:26:39.775095 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4386b4f2-7580-4c18-ae6a-078f8de0edc5/glance-httpd/0.log" Dec 02 11:26:39 crc kubenswrapper[4679]: I1202 11:26:39.778931 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4386b4f2-7580-4c18-ae6a-078f8de0edc5/glance-log/0.log" Dec 02 11:26:39 crc kubenswrapper[4679]: I1202 11:26:39.864863 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_daa65ea1-139c-49c6-9f88-36efbb79d52c/glance-httpd/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.010548 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_daa65ea1-139c-49c6-9f88-36efbb79d52c/glance-log/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.143117 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6fd99d7d5b-sb96h_4dd62da9-d481-4fc3-bbe6-108a54d98154/horizon/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.321326 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-2wkpm_ca2fc1cb-e547-47c7-b27a-7de8429d2032/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.405919 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-nr9dr_645181f0-3043-4bad-908e-8d57baa1d9b9/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.468829 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6fd99d7d5b-sb96h_4dd62da9-d481-4fc3-bbe6-108a54d98154/horizon-log/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.622984 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29411221-9cggb_dea51cbc-b2de-4d00-a13e-fb2e1c66b605/keystone-cron/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.728744 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6558c68964-dhdww_6691f3fc-9b58-4cc5-af4e-b4aaf10036e9/keystone-api/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.831812 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_84bc340a-3e0b-4ac5-94a7-07156dbcbbf0/kube-state-metrics/0.log" Dec 02 11:26:40 crc kubenswrapper[4679]: I1202 11:26:40.907896 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lhf76_1202831e-ebbe-4507-b89a-b3e9de9d64e6/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:41 crc kubenswrapper[4679]: I1202 11:26:41.204762 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6dbdc9c48f-dtqv6_eecb1bea-c78b-4888-b9a8-e868ab82a9db/neutron-api/0.log" Dec 02 11:26:41 crc kubenswrapper[4679]: I1202 11:26:41.252150 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6dbdc9c48f-dtqv6_eecb1bea-c78b-4888-b9a8-e868ab82a9db/neutron-httpd/0.log" Dec 02 11:26:41 crc kubenswrapper[4679]: I1202 11:26:41.375117 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-r872r_7764127f-88d3-4c9e-9eb2-9a6c37585ca3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:41 crc kubenswrapper[4679]: I1202 11:26:41.928032 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_81f22355-2ad7-4fd6-bc88-779cd704b038/nova-api-log/0.log" Dec 02 11:26:41 crc kubenswrapper[4679]: I1202 11:26:41.959636 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_759b2c49-d8b4-4f12-b60d-3ad7765f8aa3/nova-cell0-conductor-conductor/0.log" Dec 02 11:26:42 crc kubenswrapper[4679]: I1202 11:26:42.346026 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_95edebcc-4231-4f68-be9d-7e04a0d15e16/nova-cell1-conductor-conductor/0.log" Dec 02 11:26:42 crc kubenswrapper[4679]: I1202 11:26:42.359345 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_81f22355-2ad7-4fd6-bc88-779cd704b038/nova-api-api/0.log" Dec 02 11:26:42 crc kubenswrapper[4679]: I1202 11:26:42.360866 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_30a95365-228b-4db8-bcae-4957f4a4ee87/nova-cell1-novncproxy-novncproxy/0.log" Dec 02 11:26:42 crc kubenswrapper[4679]: I1202 11:26:42.605280 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-nvlmk_eae6d050-4b37-4c45-a0bd-4b930852eecf/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:42 crc kubenswrapper[4679]: I1202 11:26:42.686692 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8d049f08-8029-4ef7-9383-34925ba3d48b/nova-metadata-log/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.061237 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d731d782-b47c-4e03-8191-47ae5d52a555/nova-scheduler-scheduler/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.104374 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_30ab65a3-79d4-4eb7-a6b3-c759422ddffc/mysql-bootstrap/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.320280 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_30ab65a3-79d4-4eb7-a6b3-c759422ddffc/mysql-bootstrap/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.347560 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_30ab65a3-79d4-4eb7-a6b3-c759422ddffc/galera/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.520418 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e1f514bf-70bf-484c-b485-7e3d46a5c139/mysql-bootstrap/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.742031 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e1f514bf-70bf-484c-b485-7e3d46a5c139/galera/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.754003 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e1f514bf-70bf-484c-b485-7e3d46a5c139/mysql-bootstrap/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.910424 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8d049f08-8029-4ef7-9383-34925ba3d48b/nova-metadata-metadata/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.973919 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kg24f_142f095e-2fa0-478d-a611-50aad287d2c6/openstack-network-exporter/0.log" Dec 02 11:26:43 crc kubenswrapper[4679]: I1202 11:26:43.980209 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c0d970b7-84af-4667-8b4e-6b2d4b800923/openstackclient/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.157941 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovsdb-server-init/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.333121 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovs-vswitchd/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.345428 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovsdb-server/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.350623 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z6rx5_9a9d4b4d-2f9c-4f27-9f64-98efad54fded/ovsdb-server-init/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.526419 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p5dsw_9ed5ae26-6384-4c09-9fa3-c9f8dd7fa9bb/ovn-controller/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.617354 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gzdgr_7e6fa9dc-0546-45fd-847c-29266ce64dfa/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.720698 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ae2e8d7b-ca00-404b-a11b-4ea010116548/openstack-network-exporter/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.820690 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ae2e8d7b-ca00-404b-a11b-4ea010116548/ovn-northd/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.944348 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_71d9df94-9e39-409b-aa9e-e13a8f8f9ea2/openstack-network-exporter/0.log" Dec 02 11:26:44 crc kubenswrapper[4679]: I1202 11:26:44.964949 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_71d9df94-9e39-409b-aa9e-e13a8f8f9ea2/ovsdbserver-nb/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.271882 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_285ae6d4-044a-4720-a495-8ca9eca431f3/openstack-network-exporter/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.357124 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_285ae6d4-044a-4720-a495-8ca9eca431f3/ovsdbserver-sb/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.594208 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17e32beb-4297-4540-a48b-b1d3d4a32ee2/setup-container/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.609676 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-947c4b5c6-7fq4t_19f15bf3-d397-4937-bcd3-c61999ce8dae/placement-api/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.636387 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-947c4b5c6-7fq4t_19f15bf3-d397-4937-bcd3-c61999ce8dae/placement-log/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.824736 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_561abb24-2326-4992-b863-6065fbdcb959/setup-container/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.825158 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17e32beb-4297-4540-a48b-b1d3d4a32ee2/setup-container/0.log" Dec 02 11:26:45 crc kubenswrapper[4679]: I1202 11:26:45.851116 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17e32beb-4297-4540-a48b-b1d3d4a32ee2/rabbitmq/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.087849 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_561abb24-2326-4992-b863-6065fbdcb959/setup-container/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.116231 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_561abb24-2326-4992-b863-6065fbdcb959/rabbitmq/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.189490 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-mfdrj_0430fbc5-3d73-4885-8670-61dd256714be/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.331943 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-bzv4b_e5d79c8f-5612-4597-ba1a-51108bf84927/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.444150 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-xp2m6_52ca33f5-24c7-4dfb-ad4f-6e2b67c1849a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.588237 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-7dlsf_2fd59ebd-72da-4ad7-93b5-7f51f3a3304c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.665607 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-jtmsz_8ff3dc59-0c1e-49c2-a459-7c0dd461f9d7/ssh-known-hosts-edpm-deployment/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.886185 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8d8658b7-4bkfp_8d434783-1fe8-4ba4-8195-282fa213066d/proxy-server/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.944782 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8d8658b7-4bkfp_8d434783-1fe8-4ba4-8195-282fa213066d/proxy-httpd/0.log" Dec 02 11:26:46 crc kubenswrapper[4679]: I1202 11:26:46.995766 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-2wk8m_714d01a2-8f79-41d8-a16a-47c35a69fe8b/swift-ring-rebalance/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.154171 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-auditor/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.171427 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-reaper/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.283856 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-replicator/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.638108 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/account-server/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.690529 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-auditor/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.700539 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-server/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.703260 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-replicator/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.816092 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/container-updater/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.893425 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-auditor/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.908543 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:26:47 crc kubenswrapper[4679]: E1202 11:26:47.908788 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.929815 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-expirer/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.954177 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-replicator/0.log" Dec 02 11:26:47 crc kubenswrapper[4679]: I1202 11:26:47.995396 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-server/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.125473 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/object-updater/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.144534 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/swift-recon-cron/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.149115 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d8428a26-75d9-4d87-b31d-79329264e14e/rsync/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.377566 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-hwmw6_d79ba6bd-2e41-485d-a11f-5b0bdb5500c8/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.421488 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_564a5130-acc1-40e0-814b-2e9bfd2d967d/tempest-tests-tempest-tests-runner/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.582565 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d17cc270-8f47-4ea2-ba89-6bb3011ed620/test-operator-logs-container/0.log" Dec 02 11:26:48 crc kubenswrapper[4679]: I1202 11:26:48.684124 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5nrms_1420f674-925e-449b-8e18-0e97225f3958/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 11:26:58 crc kubenswrapper[4679]: I1202 11:26:58.908644 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:26:58 crc kubenswrapper[4679]: E1202 11:26:58.909418 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:27:01 crc kubenswrapper[4679]: I1202 11:27:01.634610 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_afdf0aae-7bec-4c47-820f-5a970c264069/memcached/0.log" Dec 02 11:27:12 crc kubenswrapper[4679]: I1202 11:27:12.909769 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:27:12 crc kubenswrapper[4679]: E1202 11:27:12.910706 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.248339 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jmj2d_48f2b621-5d9b-4213-8681-8107d1fc04a4/kube-rbac-proxy/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.350084 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jmj2d_48f2b621-5d9b-4213-8681-8107d1fc04a4/manager/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.455907 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-snss6_7938c8f2-53f6-4143-997d-a44ea54c9fc2/kube-rbac-proxy/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.516856 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-snss6_7938c8f2-53f6-4143-997d-a44ea54c9fc2/manager/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.640883 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/util/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.772575 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/util/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.811890 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/pull/0.log" Dec 02 11:27:16 crc kubenswrapper[4679]: I1202 11:27:16.819527 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/pull/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.443011 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/util/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.470671 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/pull/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.470905 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65ad8q44_f5705dac-dfbc-4996-b0b2-177ce907b32a/extract/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.624411 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-59k6f_85ed4b5c-eda1-481a-84ae-b97d01a9b9eb/kube-rbac-proxy/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.644032 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-59k6f_85ed4b5c-eda1-481a-84ae-b97d01a9b9eb/manager/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.669029 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-442vr_61bf694c-dbb6-442c-aaa0-3c13ff7efc97/kube-rbac-proxy/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.861609 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-442vr_61bf694c-dbb6-442c-aaa0-3c13ff7efc97/manager/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.876373 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jxm5w_f9cf4107-9e23-4246-a8b7-b3bbda0cc461/kube-rbac-proxy/0.log" Dec 02 11:27:17 crc kubenswrapper[4679]: I1202 11:27:17.951932 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jxm5w_f9cf4107-9e23-4246-a8b7-b3bbda0cc461/manager/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.044194 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bnz2w_80f86afb-0d79-4d1c-a0fd-b1d573d3c471/kube-rbac-proxy/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.084686 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bnz2w_80f86afb-0d79-4d1c-a0fd-b1d573d3c471/manager/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.202004 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-ghgsb_a0088242-8397-41e7-96b8-d971f307eefc/kube-rbac-proxy/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.374218 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-ghgsb_a0088242-8397-41e7-96b8-d971f307eefc/manager/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.403608 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-5zzmd_5ee03695-d4bb-4b77-84d1-84190ae307d6/kube-rbac-proxy/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.424987 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-5zzmd_5ee03695-d4bb-4b77-84d1-84190ae307d6/manager/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.604580 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-sxgbw_ebbb7257-823e-49a1-af4c-703b979de700/kube-rbac-proxy/0.log" Dec 02 11:27:18 crc kubenswrapper[4679]: I1202 11:27:18.655338 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-sxgbw_ebbb7257-823e-49a1-af4c-703b979de700/manager/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.161003 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-w74h9_e334fbef-5aab-4cd6-83a1-084160e32d90/kube-rbac-proxy/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.171347 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-tp7zw_1cf6ea83-714b-4a9d-8a5e-708293a3e9e6/kube-rbac-proxy/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.197457 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-tp7zw_1cf6ea83-714b-4a9d-8a5e-708293a3e9e6/manager/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.341220 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-w74h9_e334fbef-5aab-4cd6-83a1-084160e32d90/manager/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.408367 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-48bsh_e23de616-947b-4ff1-b05b-522a1c39cc7b/kube-rbac-proxy/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.423074 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-48bsh_e23de616-947b-4ff1-b05b-522a1c39cc7b/manager/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.546410 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-628rp_9dc9ab59-e368-452c-9ce2-e4d028fd35a1/kube-rbac-proxy/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.681619 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-628rp_9dc9ab59-e368-452c-9ce2-e4d028fd35a1/manager/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.748107 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-nm552_1b3b0e6e-eb2f-40fc-a77f-c8889816dee3/kube-rbac-proxy/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.810927 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-nm552_1b3b0e6e-eb2f-40fc-a77f-c8889816dee3/manager/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.924090 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh_b255bbc9-0452-4445-9470-8bd7ad85257a/kube-rbac-proxy/0.log" Dec 02 11:27:19 crc kubenswrapper[4679]: I1202 11:27:19.925038 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4fv4gh_b255bbc9-0452-4445-9470-8bd7ad85257a/manager/0.log" Dec 02 11:27:20 crc kubenswrapper[4679]: I1202 11:27:20.452355 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wh4rf_7fa3fa68-e237-45d4-b1b4-7422e84f46ef/registry-server/0.log" Dec 02 11:27:20 crc kubenswrapper[4679]: I1202 11:27:20.496195 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-655b495c79-ktwld_7b731918-9630-4abe-a4ce-ce8a65167295/operator/0.log" Dec 02 11:27:20 crc kubenswrapper[4679]: I1202 11:27:20.584435 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xfz4c_3fa279c2-e6ef-4131-ab48-28306a24000e/kube-rbac-proxy/0.log" Dec 02 11:27:20 crc kubenswrapper[4679]: I1202 11:27:20.746193 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xfz4c_3fa279c2-e6ef-4131-ab48-28306a24000e/manager/0.log" Dec 02 11:27:20 crc kubenswrapper[4679]: I1202 11:27:20.752129 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9flx2_c26eed77-aa87-4b6b-9f81-aefe3e5ed786/kube-rbac-proxy/0.log" Dec 02 11:27:20 crc kubenswrapper[4679]: I1202 11:27:20.858850 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-9flx2_c26eed77-aa87-4b6b-9f81-aefe3e5ed786/manager/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.016717 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lxjdn_0a119e38-647a-42ab-823f-e5ecea4afbf0/operator/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.021364 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7bc5fbbb46-rqv9x_6edbf2ab-201d-4a7f-902a-7b74566b5a36/manager/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.124409 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nw7fm_a877b082-6f7d-4f85-b290-356fe7c8bb5f/kube-rbac-proxy/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.185695 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nw7fm_a877b082-6f7d-4f85-b290-356fe7c8bb5f/manager/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.208807 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-ps9nb_921e63bc-e66b-48d5-a22c-010c65a0e1b0/kube-rbac-proxy/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.317256 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-ps9nb_921e63bc-e66b-48d5-a22c-010c65a0e1b0/manager/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.377472 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-b74hj_6ac36ac3-f2ba-40be-9544-9897a27ddaf4/kube-rbac-proxy/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.422501 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-b74hj_6ac36ac3-f2ba-40be-9544-9897a27ddaf4/manager/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.508979 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7kl62_f48fce46-9ae8-4154-b67b-3edd56267209/kube-rbac-proxy/0.log" Dec 02 11:27:21 crc kubenswrapper[4679]: I1202 11:27:21.578029 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-7kl62_f48fce46-9ae8-4154-b67b-3edd56267209/manager/0.log" Dec 02 11:27:27 crc kubenswrapper[4679]: I1202 11:27:27.199891 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:27:27 crc kubenswrapper[4679]: E1202 11:27:27.200574 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:27:37 crc kubenswrapper[4679]: I1202 11:27:37.908780 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:27:37 crc kubenswrapper[4679]: E1202 11:27:37.909582 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:27:41 crc kubenswrapper[4679]: I1202 11:27:41.174775 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ffc27_6e31c209-d885-4235-87a7-972b045a6c72/control-plane-machine-set-operator/0.log" Dec 02 11:27:41 crc kubenswrapper[4679]: I1202 11:27:41.235249 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z95rb_e222153a-8cab-4ef0-8227-13be178f3f82/kube-rbac-proxy/0.log" Dec 02 11:27:41 crc kubenswrapper[4679]: I1202 11:27:41.364341 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-z95rb_e222153a-8cab-4ef0-8227-13be178f3f82/machine-api-operator/0.log" Dec 02 11:27:52 crc kubenswrapper[4679]: I1202 11:27:52.908982 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:27:52 crc kubenswrapper[4679]: E1202 11:27:52.909954 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:27:53 crc kubenswrapper[4679]: I1202 11:27:53.899220 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-rnwm9_1da80a36-7d00-4055-ba3e-1e84a58e2748/cert-manager-controller/0.log" Dec 02 11:27:54 crc kubenswrapper[4679]: I1202 11:27:54.074644 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dhkj7_227ec57c-bcbe-4725-8726-3411f96b09ac/cert-manager-cainjector/0.log" Dec 02 11:27:54 crc kubenswrapper[4679]: I1202 11:27:54.095844 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-lhvkg_f91be00f-75e8-4a8b-aacc-d881b6af57a8/cert-manager-webhook/0.log" Dec 02 11:28:06 crc kubenswrapper[4679]: I1202 11:28:06.229481 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-r2xnw_932b0be3-5659-4b5b-a667-cd03f56c8f98/nmstate-console-plugin/0.log" Dec 02 11:28:06 crc kubenswrapper[4679]: I1202 11:28:06.433992 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-t6qm2_49ed79ab-5582-4f7e-bfd9-18263fa9ccc2/nmstate-handler/0.log" Dec 02 11:28:06 crc kubenswrapper[4679]: I1202 11:28:06.454754 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9w2kz_1daefa02-7428-44b6-8da5-f1ecf0d5c12e/kube-rbac-proxy/0.log" Dec 02 11:28:06 crc kubenswrapper[4679]: I1202 11:28:06.502403 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9w2kz_1daefa02-7428-44b6-8da5-f1ecf0d5c12e/nmstate-metrics/0.log" Dec 02 11:28:06 crc kubenswrapper[4679]: I1202 11:28:06.649791 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-hnqnh_0cdf81d6-53ba-4fae-95ff-6d0d28b2384e/nmstate-operator/0.log" Dec 02 11:28:06 crc kubenswrapper[4679]: I1202 11:28:06.722935 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-fmbsj_338504ed-81c0-4c88-8d99-b1ffc9cd7bcb/nmstate-webhook/0.log" Dec 02 11:28:07 crc kubenswrapper[4679]: I1202 11:28:07.908998 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:28:07 crc kubenswrapper[4679]: E1202 11:28:07.909665 4679 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lzf8q_openshift-machine-config-operator(5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb)\"" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.114636 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-ws8bf_d2302ab5-3f22-4f2e-804f-a9e1109446a6/kube-rbac-proxy/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.261906 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-ws8bf_d2302ab5-3f22-4f2e-804f-a9e1109446a6/controller/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.306164 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.535378 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.535401 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.548072 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.577129 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:28:21 crc kubenswrapper[4679]: I1202 11:28:21.909225 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.046553 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.084035 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.106289 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.113408 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.253487 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-reloader/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.266695 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-frr-files/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.284669 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/cp-metrics/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.292763 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/controller/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.471934 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/frr-metrics/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.477783 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/kube-rbac-proxy-frr/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.483357 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/kube-rbac-proxy/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.681709 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-zcqd5_280f2b1e-4e5a-481d-ad2f-4536d832697a/frr-k8s-webhook-server/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.696280 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/reloader/0.log" Dec 02 11:28:22 crc kubenswrapper[4679]: I1202 11:28:22.904655 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8555c5bdfd-j4hqj_06025ad8-fd55-4fe1-95bd-08a0c83168e0/manager/0.log" Dec 02 11:28:23 crc kubenswrapper[4679]: I1202 11:28:23.099119 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-898774cd7-nvg9q_4086bfe3-e82b-46c4-87f6-837a42348472/webhook-server/0.log" Dec 02 11:28:23 crc kubenswrapper[4679]: I1202 11:28:23.099560 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"631ae3e93387fca9569a2e844307a8cec9d81cf13cf544c4d8fad7c5a98ae3cc"} Dec 02 11:28:23 crc kubenswrapper[4679]: I1202 11:28:23.153459 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qgdwn_c42448eb-c465-4172-9830-37000ff7e881/kube-rbac-proxy/0.log" Dec 02 11:28:23 crc kubenswrapper[4679]: I1202 11:28:23.747581 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qgdwn_c42448eb-c465-4172-9830-37000ff7e881/speaker/0.log" Dec 02 11:28:23 crc kubenswrapper[4679]: I1202 11:28:23.955775 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-qx64n_c5c9e3af-4021-4100-8a4f-3523f5da7bf9/frr/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.191985 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/util/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.408578 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/pull/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.461245 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/util/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.494217 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/pull/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.593611 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/util/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.599980 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/pull/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.620489 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212ftdt86_ae8d2b3f-baec-4d47-be99-1d73bc31416d/extract/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.738736 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/util/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.921460 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/util/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.933655 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/pull/0.log" Dec 02 11:28:36 crc kubenswrapper[4679]: I1202 11:28:36.966438 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/pull/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.097920 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/util/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.112756 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/pull/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.124776 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h4rtm_646ec543-e35d-4099-bf1a-d0284276fc95/extract/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.276211 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-utilities/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.433194 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-utilities/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.433572 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-content/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.456318 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-content/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.601372 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-content/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.613646 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/extract-utilities/0.log" Dec 02 11:28:37 crc kubenswrapper[4679]: I1202 11:28:37.810281 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-utilities/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.037295 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dk6hk_33efe285-5788-489a-98b5-c6cce7f2738c/registry-server/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.042881 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-utilities/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.060201 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-content/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.074020 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-content/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.205508 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-content/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.226512 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/extract-utilities/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.453586 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-72flf_de76d395-f3a6-474d-8ab7-f31c46710a0d/marketplace-operator/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.556610 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-utilities/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.559154 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-b2xjv_41a03b9a-9119-474c-9ae4-725879cbc78e/registry-server/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.687432 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-utilities/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.708903 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-content/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.747651 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-content/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.886974 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-content/0.log" Dec 02 11:28:38 crc kubenswrapper[4679]: I1202 11:28:38.921851 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/extract-utilities/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.037422 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dg7kk_1f4b93fd-d2b3-4f28-aff5-1a4dd8d79cba/registry-server/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.094023 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-utilities/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.246706 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-content/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.278035 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-utilities/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.281802 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-content/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.431776 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-utilities/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.466555 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/extract-content/0.log" Dec 02 11:28:39 crc kubenswrapper[4679]: I1202 11:28:39.965835 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z6xb5_e534b0e4-8964-48d5-8bad-734ec5d93a8b/registry-server/0.log" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.648132 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k7lv9"] Dec 02 11:28:45 crc kubenswrapper[4679]: E1202 11:28:45.649025 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6009500f-55a2-4026-8347-9e77410c710c" containerName="container-00" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.649040 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="6009500f-55a2-4026-8347-9e77410c710c" containerName="container-00" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.649333 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="6009500f-55a2-4026-8347-9e77410c710c" containerName="container-00" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.651789 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.662693 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k7lv9"] Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.703498 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c4v9\" (UniqueName: \"kubernetes.io/projected/f4291552-9aa3-42aa-b985-270c84fa21cb-kube-api-access-6c4v9\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.703651 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-catalog-content\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.703712 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-utilities\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.805835 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c4v9\" (UniqueName: \"kubernetes.io/projected/f4291552-9aa3-42aa-b985-270c84fa21cb-kube-api-access-6c4v9\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.805971 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-catalog-content\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.806025 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-utilities\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.806911 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-catalog-content\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.807007 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-utilities\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.828620 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c4v9\" (UniqueName: \"kubernetes.io/projected/f4291552-9aa3-42aa-b985-270c84fa21cb-kube-api-access-6c4v9\") pod \"certified-operators-k7lv9\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:45 crc kubenswrapper[4679]: I1202 11:28:45.984781 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:46 crc kubenswrapper[4679]: I1202 11:28:46.487848 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k7lv9"] Dec 02 11:28:47 crc kubenswrapper[4679]: I1202 11:28:47.325282 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerID="16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190" exitCode=0 Dec 02 11:28:47 crc kubenswrapper[4679]: I1202 11:28:47.325337 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerDied","Data":"16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190"} Dec 02 11:28:47 crc kubenswrapper[4679]: I1202 11:28:47.325381 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerStarted","Data":"29065aa0b43eff43c11639e906207d5bbf62ae5f058e6ebcf8f08845dbc49532"} Dec 02 11:28:47 crc kubenswrapper[4679]: I1202 11:28:47.327518 4679 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 11:28:48 crc kubenswrapper[4679]: I1202 11:28:48.335380 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerStarted","Data":"0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381"} Dec 02 11:28:49 crc kubenswrapper[4679]: I1202 11:28:49.347451 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerID="0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381" exitCode=0 Dec 02 11:28:49 crc kubenswrapper[4679]: I1202 11:28:49.347511 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerDied","Data":"0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381"} Dec 02 11:28:50 crc kubenswrapper[4679]: I1202 11:28:50.362398 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerStarted","Data":"f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e"} Dec 02 11:28:50 crc kubenswrapper[4679]: I1202 11:28:50.391636 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k7lv9" podStartSLOduration=2.882048407 podStartE2EDuration="5.391617782s" podCreationTimestamp="2025-12-02 11:28:45 +0000 UTC" firstStartedPulling="2025-12-02 11:28:47.327271332 +0000 UTC m=+4180.657410192" lastFinishedPulling="2025-12-02 11:28:49.836840707 +0000 UTC m=+4183.166979567" observedRunningTime="2025-12-02 11:28:50.384622893 +0000 UTC m=+4183.714761753" watchObservedRunningTime="2025-12-02 11:28:50.391617782 +0000 UTC m=+4183.721756642" Dec 02 11:28:55 crc kubenswrapper[4679]: I1202 11:28:55.985983 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:55 crc kubenswrapper[4679]: I1202 11:28:55.986654 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:56 crc kubenswrapper[4679]: I1202 11:28:56.033422 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:56 crc kubenswrapper[4679]: I1202 11:28:56.464701 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:56 crc kubenswrapper[4679]: I1202 11:28:56.515117 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k7lv9"] Dec 02 11:28:58 crc kubenswrapper[4679]: I1202 11:28:58.431181 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k7lv9" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="registry-server" containerID="cri-o://f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e" gracePeriod=2 Dec 02 11:28:58 crc kubenswrapper[4679]: I1202 11:28:58.949977 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.062097 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-utilities\") pod \"f4291552-9aa3-42aa-b985-270c84fa21cb\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.062237 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-catalog-content\") pod \"f4291552-9aa3-42aa-b985-270c84fa21cb\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.062292 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c4v9\" (UniqueName: \"kubernetes.io/projected/f4291552-9aa3-42aa-b985-270c84fa21cb-kube-api-access-6c4v9\") pod \"f4291552-9aa3-42aa-b985-270c84fa21cb\" (UID: \"f4291552-9aa3-42aa-b985-270c84fa21cb\") " Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.062912 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-utilities" (OuterVolumeSpecName: "utilities") pod "f4291552-9aa3-42aa-b985-270c84fa21cb" (UID: "f4291552-9aa3-42aa-b985-270c84fa21cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.076533 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4291552-9aa3-42aa-b985-270c84fa21cb-kube-api-access-6c4v9" (OuterVolumeSpecName: "kube-api-access-6c4v9") pod "f4291552-9aa3-42aa-b985-270c84fa21cb" (UID: "f4291552-9aa3-42aa-b985-270c84fa21cb"). InnerVolumeSpecName "kube-api-access-6c4v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.111762 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4291552-9aa3-42aa-b985-270c84fa21cb" (UID: "f4291552-9aa3-42aa-b985-270c84fa21cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.164781 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.164833 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c4v9\" (UniqueName: \"kubernetes.io/projected/f4291552-9aa3-42aa-b985-270c84fa21cb-kube-api-access-6c4v9\") on node \"crc\" DevicePath \"\"" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.164850 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4291552-9aa3-42aa-b985-270c84fa21cb-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.445265 4679 generic.go:334] "Generic (PLEG): container finished" podID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerID="f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e" exitCode=0 Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.445357 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerDied","Data":"f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e"} Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.445384 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k7lv9" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.445705 4679 scope.go:117] "RemoveContainer" containerID="f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.445691 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k7lv9" event={"ID":"f4291552-9aa3-42aa-b985-270c84fa21cb","Type":"ContainerDied","Data":"29065aa0b43eff43c11639e906207d5bbf62ae5f058e6ebcf8f08845dbc49532"} Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.486242 4679 scope.go:117] "RemoveContainer" containerID="0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.509778 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k7lv9"] Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.518112 4679 scope.go:117] "RemoveContainer" containerID="16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.525998 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k7lv9"] Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.552472 4679 scope.go:117] "RemoveContainer" containerID="f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e" Dec 02 11:28:59 crc kubenswrapper[4679]: E1202 11:28:59.552990 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e\": container with ID starting with f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e not found: ID does not exist" containerID="f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.553025 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e"} err="failed to get container status \"f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e\": rpc error: code = NotFound desc = could not find container \"f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e\": container with ID starting with f4b608ec7f34d887b1459f4d89d8a6350f2b2df58eca419a93cb7b66f207e58e not found: ID does not exist" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.553070 4679 scope.go:117] "RemoveContainer" containerID="0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381" Dec 02 11:28:59 crc kubenswrapper[4679]: E1202 11:28:59.561679 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381\": container with ID starting with 0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381 not found: ID does not exist" containerID="0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.561728 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381"} err="failed to get container status \"0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381\": rpc error: code = NotFound desc = could not find container \"0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381\": container with ID starting with 0136ee39184668ee5def60db85f7be5f7bfd86cd12de0f647df87a83e3d21381 not found: ID does not exist" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.561759 4679 scope.go:117] "RemoveContainer" containerID="16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190" Dec 02 11:28:59 crc kubenswrapper[4679]: E1202 11:28:59.562157 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190\": container with ID starting with 16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190 not found: ID does not exist" containerID="16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190" Dec 02 11:28:59 crc kubenswrapper[4679]: I1202 11:28:59.562183 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190"} err="failed to get container status \"16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190\": rpc error: code = NotFound desc = could not find container \"16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190\": container with ID starting with 16bd0503e6541bdef395782711bb9a94c6c80b8fcdf1e245a906ddcd8e64d190 not found: ID does not exist" Dec 02 11:29:00 crc kubenswrapper[4679]: I1202 11:29:00.920669 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" path="/var/lib/kubelet/pods/f4291552-9aa3-42aa-b985-270c84fa21cb/volumes" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.451571 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-glm58"] Dec 02 11:29:50 crc kubenswrapper[4679]: E1202 11:29:50.453899 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="extract-content" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.453993 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="extract-content" Dec 02 11:29:50 crc kubenswrapper[4679]: E1202 11:29:50.454061 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="registry-server" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.454122 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="registry-server" Dec 02 11:29:50 crc kubenswrapper[4679]: E1202 11:29:50.454216 4679 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="extract-utilities" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.454280 4679 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="extract-utilities" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.454591 4679 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4291552-9aa3-42aa-b985-270c84fa21cb" containerName="registry-server" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.456048 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.480158 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glm58"] Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.590244 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-catalog-content\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.590452 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-utilities\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.590519 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfffj\" (UniqueName: \"kubernetes.io/projected/6c692314-e421-4ee8-aa31-6ea752ba80ce-kube-api-access-vfffj\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.692160 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-utilities\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.692267 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfffj\" (UniqueName: \"kubernetes.io/projected/6c692314-e421-4ee8-aa31-6ea752ba80ce-kube-api-access-vfffj\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.692327 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-catalog-content\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.692775 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-utilities\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.692797 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-catalog-content\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:50 crc kubenswrapper[4679]: I1202 11:29:50.955687 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfffj\" (UniqueName: \"kubernetes.io/projected/6c692314-e421-4ee8-aa31-6ea752ba80ce-kube-api-access-vfffj\") pod \"community-operators-glm58\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:51 crc kubenswrapper[4679]: I1202 11:29:51.080670 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:29:51 crc kubenswrapper[4679]: I1202 11:29:51.577389 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glm58"] Dec 02 11:29:51 crc kubenswrapper[4679]: W1202 11:29:51.590695 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c692314_e421_4ee8_aa31_6ea752ba80ce.slice/crio-56a0b5577de9f4c5dd7830204f3a76c2b15da746a9b04195300c611d3b8be326 WatchSource:0}: Error finding container 56a0b5577de9f4c5dd7830204f3a76c2b15da746a9b04195300c611d3b8be326: Status 404 returned error can't find the container with id 56a0b5577de9f4c5dd7830204f3a76c2b15da746a9b04195300c611d3b8be326 Dec 02 11:29:52 crc kubenswrapper[4679]: I1202 11:29:52.100696 4679 generic.go:334] "Generic (PLEG): container finished" podID="6c692314-e421-4ee8-aa31-6ea752ba80ce" containerID="528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23" exitCode=0 Dec 02 11:29:52 crc kubenswrapper[4679]: I1202 11:29:52.100799 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerDied","Data":"528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23"} Dec 02 11:29:52 crc kubenswrapper[4679]: I1202 11:29:52.100831 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerStarted","Data":"56a0b5577de9f4c5dd7830204f3a76c2b15da746a9b04195300c611d3b8be326"} Dec 02 11:29:54 crc kubenswrapper[4679]: I1202 11:29:54.124100 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerStarted","Data":"8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59"} Dec 02 11:29:55 crc kubenswrapper[4679]: I1202 11:29:55.141377 4679 generic.go:334] "Generic (PLEG): container finished" podID="6c692314-e421-4ee8-aa31-6ea752ba80ce" containerID="8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59" exitCode=0 Dec 02 11:29:55 crc kubenswrapper[4679]: I1202 11:29:55.141416 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerDied","Data":"8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59"} Dec 02 11:29:56 crc kubenswrapper[4679]: I1202 11:29:56.154935 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerStarted","Data":"3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373"} Dec 02 11:29:56 crc kubenswrapper[4679]: I1202 11:29:56.180477 4679 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-glm58" podStartSLOduration=2.454366544 podStartE2EDuration="6.180457273s" podCreationTimestamp="2025-12-02 11:29:50 +0000 UTC" firstStartedPulling="2025-12-02 11:29:52.107021601 +0000 UTC m=+4245.437160501" lastFinishedPulling="2025-12-02 11:29:55.83311238 +0000 UTC m=+4249.163251230" observedRunningTime="2025-12-02 11:29:56.17859342 +0000 UTC m=+4249.508732300" watchObservedRunningTime="2025-12-02 11:29:56.180457273 +0000 UTC m=+4249.510596153" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.187038 4679 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76"] Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.188438 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.190800 4679 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.196986 4679 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.215002 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76"] Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.313543 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfwfp\" (UniqueName: \"kubernetes.io/projected/a8841ac7-eae4-4095-b633-9436b53059f3-kube-api-access-gfwfp\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.313656 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8841ac7-eae4-4095-b633-9436b53059f3-config-volume\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.313716 4679 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8841ac7-eae4-4095-b633-9436b53059f3-secret-volume\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.414956 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfwfp\" (UniqueName: \"kubernetes.io/projected/a8841ac7-eae4-4095-b633-9436b53059f3-kube-api-access-gfwfp\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.415064 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8841ac7-eae4-4095-b633-9436b53059f3-config-volume\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.415126 4679 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8841ac7-eae4-4095-b633-9436b53059f3-secret-volume\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.416361 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8841ac7-eae4-4095-b633-9436b53059f3-config-volume\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.428425 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8841ac7-eae4-4095-b633-9436b53059f3-secret-volume\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.440110 4679 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfwfp\" (UniqueName: \"kubernetes.io/projected/a8841ac7-eae4-4095-b633-9436b53059f3-kube-api-access-gfwfp\") pod \"collect-profiles-29411250-mcl76\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:00 crc kubenswrapper[4679]: I1202 11:30:00.511693 4679 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:01 crc kubenswrapper[4679]: I1202 11:30:01.081399 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:30:01 crc kubenswrapper[4679]: I1202 11:30:01.081875 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:30:01 crc kubenswrapper[4679]: I1202 11:30:01.155187 4679 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:30:01 crc kubenswrapper[4679]: I1202 11:30:01.257220 4679 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:30:01 crc kubenswrapper[4679]: I1202 11:30:01.397853 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glm58"] Dec 02 11:30:01 crc kubenswrapper[4679]: I1202 11:30:01.533605 4679 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76"] Dec 02 11:30:01 crc kubenswrapper[4679]: W1202 11:30:01.544150 4679 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8841ac7_eae4_4095_b633_9436b53059f3.slice/crio-b9147c51a4eff475284e1aff814c1b0b8da5ca3deefb5e5cbcbe2c4f1e27475e WatchSource:0}: Error finding container b9147c51a4eff475284e1aff814c1b0b8da5ca3deefb5e5cbcbe2c4f1e27475e: Status 404 returned error can't find the container with id b9147c51a4eff475284e1aff814c1b0b8da5ca3deefb5e5cbcbe2c4f1e27475e Dec 02 11:30:02 crc kubenswrapper[4679]: I1202 11:30:02.210057 4679 generic.go:334] "Generic (PLEG): container finished" podID="a8841ac7-eae4-4095-b633-9436b53059f3" containerID="b36a7ada9194432e3e03787693da71f728c013cb8f254f6d74799c6304b36ac5" exitCode=0 Dec 02 11:30:02 crc kubenswrapper[4679]: I1202 11:30:02.210122 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" event={"ID":"a8841ac7-eae4-4095-b633-9436b53059f3","Type":"ContainerDied","Data":"b36a7ada9194432e3e03787693da71f728c013cb8f254f6d74799c6304b36ac5"} Dec 02 11:30:02 crc kubenswrapper[4679]: I1202 11:30:02.210467 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" event={"ID":"a8841ac7-eae4-4095-b633-9436b53059f3","Type":"ContainerStarted","Data":"b9147c51a4eff475284e1aff814c1b0b8da5ca3deefb5e5cbcbe2c4f1e27475e"} Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.220454 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-glm58" podUID="6c692314-e421-4ee8-aa31-6ea752ba80ce" containerName="registry-server" containerID="cri-o://3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373" gracePeriod=2 Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.627583 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.779447 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfwfp\" (UniqueName: \"kubernetes.io/projected/a8841ac7-eae4-4095-b633-9436b53059f3-kube-api-access-gfwfp\") pod \"a8841ac7-eae4-4095-b633-9436b53059f3\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.779561 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8841ac7-eae4-4095-b633-9436b53059f3-secret-volume\") pod \"a8841ac7-eae4-4095-b633-9436b53059f3\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.779629 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8841ac7-eae4-4095-b633-9436b53059f3-config-volume\") pod \"a8841ac7-eae4-4095-b633-9436b53059f3\" (UID: \"a8841ac7-eae4-4095-b633-9436b53059f3\") " Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.780461 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8841ac7-eae4-4095-b633-9436b53059f3-config-volume" (OuterVolumeSpecName: "config-volume") pod "a8841ac7-eae4-4095-b633-9436b53059f3" (UID: "a8841ac7-eae4-4095-b633-9436b53059f3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.780686 4679 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8841ac7-eae4-4095-b633-9436b53059f3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.785388 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8841ac7-eae4-4095-b633-9436b53059f3-kube-api-access-gfwfp" (OuterVolumeSpecName: "kube-api-access-gfwfp") pod "a8841ac7-eae4-4095-b633-9436b53059f3" (UID: "a8841ac7-eae4-4095-b633-9436b53059f3"). InnerVolumeSpecName "kube-api-access-gfwfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.789506 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8841ac7-eae4-4095-b633-9436b53059f3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a8841ac7-eae4-4095-b633-9436b53059f3" (UID: "a8841ac7-eae4-4095-b633-9436b53059f3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.858506 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.883034 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfwfp\" (UniqueName: \"kubernetes.io/projected/a8841ac7-eae4-4095-b633-9436b53059f3-kube-api-access-gfwfp\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.883064 4679 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8841ac7-eae4-4095-b633-9436b53059f3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.983772 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-catalog-content\") pod \"6c692314-e421-4ee8-aa31-6ea752ba80ce\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.983840 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfffj\" (UniqueName: \"kubernetes.io/projected/6c692314-e421-4ee8-aa31-6ea752ba80ce-kube-api-access-vfffj\") pod \"6c692314-e421-4ee8-aa31-6ea752ba80ce\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.983923 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-utilities\") pod \"6c692314-e421-4ee8-aa31-6ea752ba80ce\" (UID: \"6c692314-e421-4ee8-aa31-6ea752ba80ce\") " Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.988157 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-utilities" (OuterVolumeSpecName: "utilities") pod "6c692314-e421-4ee8-aa31-6ea752ba80ce" (UID: "6c692314-e421-4ee8-aa31-6ea752ba80ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:30:03 crc kubenswrapper[4679]: I1202 11:30:03.991008 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c692314-e421-4ee8-aa31-6ea752ba80ce-kube-api-access-vfffj" (OuterVolumeSpecName: "kube-api-access-vfffj") pod "6c692314-e421-4ee8-aa31-6ea752ba80ce" (UID: "6c692314-e421-4ee8-aa31-6ea752ba80ce"). InnerVolumeSpecName "kube-api-access-vfffj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.043534 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c692314-e421-4ee8-aa31-6ea752ba80ce" (UID: "6c692314-e421-4ee8-aa31-6ea752ba80ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.086433 4679 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.086472 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfffj\" (UniqueName: \"kubernetes.io/projected/6c692314-e421-4ee8-aa31-6ea752ba80ce-kube-api-access-vfffj\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.086484 4679 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c692314-e421-4ee8-aa31-6ea752ba80ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.232434 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" event={"ID":"a8841ac7-eae4-4095-b633-9436b53059f3","Type":"ContainerDied","Data":"b9147c51a4eff475284e1aff814c1b0b8da5ca3deefb5e5cbcbe2c4f1e27475e"} Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.232483 4679 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9147c51a4eff475284e1aff814c1b0b8da5ca3deefb5e5cbcbe2c4f1e27475e" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.232484 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411250-mcl76" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.235426 4679 generic.go:334] "Generic (PLEG): container finished" podID="6c692314-e421-4ee8-aa31-6ea752ba80ce" containerID="3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373" exitCode=0 Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.235478 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerDied","Data":"3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373"} Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.235546 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glm58" event={"ID":"6c692314-e421-4ee8-aa31-6ea752ba80ce","Type":"ContainerDied","Data":"56a0b5577de9f4c5dd7830204f3a76c2b15da746a9b04195300c611d3b8be326"} Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.235568 4679 scope.go:117] "RemoveContainer" containerID="3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.235592 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glm58" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.272153 4679 scope.go:117] "RemoveContainer" containerID="8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.284116 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glm58"] Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.294519 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-glm58"] Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.307639 4679 scope.go:117] "RemoveContainer" containerID="528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.329555 4679 scope.go:117] "RemoveContainer" containerID="3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373" Dec 02 11:30:04 crc kubenswrapper[4679]: E1202 11:30:04.330205 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373\": container with ID starting with 3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373 not found: ID does not exist" containerID="3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.330279 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373"} err="failed to get container status \"3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373\": rpc error: code = NotFound desc = could not find container \"3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373\": container with ID starting with 3da3f5ccbc6c26103179a07bfde550946cc0cd146806b4cab205968af2201373 not found: ID does not exist" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.330600 4679 scope.go:117] "RemoveContainer" containerID="8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59" Dec 02 11:30:04 crc kubenswrapper[4679]: E1202 11:30:04.331051 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59\": container with ID starting with 8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59 not found: ID does not exist" containerID="8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.331086 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59"} err="failed to get container status \"8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59\": rpc error: code = NotFound desc = could not find container \"8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59\": container with ID starting with 8af5acf1257531e7f8a65a3cebaba5545ad58df3ed751c71b94bd5391a7a5e59 not found: ID does not exist" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.331107 4679 scope.go:117] "RemoveContainer" containerID="528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23" Dec 02 11:30:04 crc kubenswrapper[4679]: E1202 11:30:04.331396 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23\": container with ID starting with 528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23 not found: ID does not exist" containerID="528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.331431 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23"} err="failed to get container status \"528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23\": rpc error: code = NotFound desc = could not find container \"528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23\": container with ID starting with 528952e77425fb0a8e19bf1592da6ba5794b708d3f4a30716e9346b2917b6e23 not found: ID does not exist" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.719589 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4"] Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.726954 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411205-vk9m4"] Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.932072 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ffd70a1-113d-434e-a0e7-701174046fff" path="/var/lib/kubelet/pods/2ffd70a1-113d-434e-a0e7-701174046fff/volumes" Dec 02 11:30:04 crc kubenswrapper[4679]: I1202 11:30:04.934632 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c692314-e421-4ee8-aa31-6ea752ba80ce" path="/var/lib/kubelet/pods/6c692314-e421-4ee8-aa31-6ea752ba80ce/volumes" Dec 02 11:30:19 crc kubenswrapper[4679]: I1202 11:30:19.420175 4679 generic.go:334] "Generic (PLEG): container finished" podID="e1d5ac99-bc19-482e-b627-b5072881dfad" containerID="721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340" exitCode=0 Dec 02 11:30:19 crc kubenswrapper[4679]: I1202 11:30:19.420289 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" event={"ID":"e1d5ac99-bc19-482e-b627-b5072881dfad","Type":"ContainerDied","Data":"721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340"} Dec 02 11:30:19 crc kubenswrapper[4679]: I1202 11:30:19.422435 4679 scope.go:117] "RemoveContainer" containerID="721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340" Dec 02 11:30:19 crc kubenswrapper[4679]: I1202 11:30:19.996524 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s9jkv_must-gather-hbkdr_e1d5ac99-bc19-482e-b627-b5072881dfad/gather/0.log" Dec 02 11:30:23 crc kubenswrapper[4679]: I1202 11:30:23.976504 4679 scope.go:117] "RemoveContainer" containerID="5539149bad511176128adca744e65383edf5fbea0ef6e075d47254a834e64153" Dec 02 11:30:29 crc kubenswrapper[4679]: I1202 11:30:29.510232 4679 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s9jkv/must-gather-hbkdr"] Dec 02 11:30:29 crc kubenswrapper[4679]: I1202 11:30:29.511334 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" podUID="e1d5ac99-bc19-482e-b627-b5072881dfad" containerName="copy" containerID="cri-o://3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea" gracePeriod=2 Dec 02 11:30:29 crc kubenswrapper[4679]: I1202 11:30:29.525849 4679 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s9jkv/must-gather-hbkdr"] Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.347535 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s9jkv_must-gather-hbkdr_e1d5ac99-bc19-482e-b627-b5072881dfad/copy/0.log" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.348222 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.529261 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6vvg\" (UniqueName: \"kubernetes.io/projected/e1d5ac99-bc19-482e-b627-b5072881dfad-kube-api-access-p6vvg\") pod \"e1d5ac99-bc19-482e-b627-b5072881dfad\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.529438 4679 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1d5ac99-bc19-482e-b627-b5072881dfad-must-gather-output\") pod \"e1d5ac99-bc19-482e-b627-b5072881dfad\" (UID: \"e1d5ac99-bc19-482e-b627-b5072881dfad\") " Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.538079 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d5ac99-bc19-482e-b627-b5072881dfad-kube-api-access-p6vvg" (OuterVolumeSpecName: "kube-api-access-p6vvg") pod "e1d5ac99-bc19-482e-b627-b5072881dfad" (UID: "e1d5ac99-bc19-482e-b627-b5072881dfad"). InnerVolumeSpecName "kube-api-access-p6vvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.548220 4679 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s9jkv_must-gather-hbkdr_e1d5ac99-bc19-482e-b627-b5072881dfad/copy/0.log" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.550433 4679 generic.go:334] "Generic (PLEG): container finished" podID="e1d5ac99-bc19-482e-b627-b5072881dfad" containerID="3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea" exitCode=143 Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.550766 4679 scope.go:117] "RemoveContainer" containerID="3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.550920 4679 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s9jkv/must-gather-hbkdr" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.612369 4679 scope.go:117] "RemoveContainer" containerID="721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.631611 4679 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6vvg\" (UniqueName: \"kubernetes.io/projected/e1d5ac99-bc19-482e-b627-b5072881dfad-kube-api-access-p6vvg\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.678022 4679 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d5ac99-bc19-482e-b627-b5072881dfad-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e1d5ac99-bc19-482e-b627-b5072881dfad" (UID: "e1d5ac99-bc19-482e-b627-b5072881dfad"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.734877 4679 scope.go:117] "RemoveContainer" containerID="3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.735748 4679 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1d5ac99-bc19-482e-b627-b5072881dfad-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 02 11:30:30 crc kubenswrapper[4679]: E1202 11:30:30.741160 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea\": container with ID starting with 3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea not found: ID does not exist" containerID="3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.741210 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea"} err="failed to get container status \"3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea\": rpc error: code = NotFound desc = could not find container \"3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea\": container with ID starting with 3954ef43d26296fe99ab4dea89bbfc21b353a31cbee82d809b3dd69fd33f5fea not found: ID does not exist" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.741258 4679 scope.go:117] "RemoveContainer" containerID="721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340" Dec 02 11:30:30 crc kubenswrapper[4679]: E1202 11:30:30.741903 4679 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340\": container with ID starting with 721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340 not found: ID does not exist" containerID="721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.741969 4679 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340"} err="failed to get container status \"721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340\": rpc error: code = NotFound desc = could not find container \"721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340\": container with ID starting with 721f720f23bc02186db10d4041e79b8f6ec911c009c93be3adfd2a9c46dc8340 not found: ID does not exist" Dec 02 11:30:30 crc kubenswrapper[4679]: I1202 11:30:30.918814 4679 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d5ac99-bc19-482e-b627-b5072881dfad" path="/var/lib/kubelet/pods/e1d5ac99-bc19-482e-b627-b5072881dfad/volumes" Dec 02 11:30:46 crc kubenswrapper[4679]: I1202 11:30:46.931762 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:30:46 crc kubenswrapper[4679]: I1202 11:30:46.932537 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:31:16 crc kubenswrapper[4679]: I1202 11:31:16.931949 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:31:16 crc kubenswrapper[4679]: I1202 11:31:16.932719 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:31:46 crc kubenswrapper[4679]: I1202 11:31:46.931248 4679 patch_prober.go:28] interesting pod/machine-config-daemon-lzf8q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 11:31:46 crc kubenswrapper[4679]: I1202 11:31:46.933025 4679 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 11:31:46 crc kubenswrapper[4679]: I1202 11:31:46.933068 4679 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" Dec 02 11:31:46 crc kubenswrapper[4679]: I1202 11:31:46.933506 4679 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"631ae3e93387fca9569a2e844307a8cec9d81cf13cf544c4d8fad7c5a98ae3cc"} pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 11:31:46 crc kubenswrapper[4679]: I1202 11:31:46.933550 4679 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" podUID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerName="machine-config-daemon" containerID="cri-o://631ae3e93387fca9569a2e844307a8cec9d81cf13cf544c4d8fad7c5a98ae3cc" gracePeriod=600 Dec 02 11:31:47 crc kubenswrapper[4679]: I1202 11:31:47.327749 4679 generic.go:334] "Generic (PLEG): container finished" podID="5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb" containerID="631ae3e93387fca9569a2e844307a8cec9d81cf13cf544c4d8fad7c5a98ae3cc" exitCode=0 Dec 02 11:31:47 crc kubenswrapper[4679]: I1202 11:31:47.327786 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerDied","Data":"631ae3e93387fca9569a2e844307a8cec9d81cf13cf544c4d8fad7c5a98ae3cc"} Dec 02 11:31:47 crc kubenswrapper[4679]: I1202 11:31:47.328137 4679 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lzf8q" event={"ID":"5df42f6a-dfd2-4c19-849c-d6ea75a1cbcb","Type":"ContainerStarted","Data":"2f8aa31241496061355052ededd9b8178800f801bfc1dfde1dc7bfe7dbf657bd"} Dec 02 11:31:47 crc kubenswrapper[4679]: I1202 11:31:47.328154 4679 scope.go:117] "RemoveContainer" containerID="47a7cee412f4d55b017c474bf9c0f5639599e402c10a5e85e64be008acf83fd8" Dec 02 11:32:24 crc kubenswrapper[4679]: I1202 11:32:24.091729 4679 scope.go:117] "RemoveContainer" containerID="48626f710891507314d5fa36e7f5096df92661247b8ab235adf117c6f678eec8" Dec 02 11:32:24 crc kubenswrapper[4679]: I1202 11:32:24.144363 4679 scope.go:117] "RemoveContainer" containerID="d92fefe3bb1f71c6d247e276764700af6f076811179648e6f61ece326f58cc0c" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113547266024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113547266017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113536126016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113536126015460 5ustar corecore